<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Integr. Neurosci.</journal-id>
<journal-title>Frontiers in Integrative Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Integr. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-5145</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnint.2023.1120668</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Working memory load modulates the processing of audiovisual distractors: A behavioral and event-related potentials study</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Yuan</surname> <given-names>Yichen</given-names></name>
<xref ref-type="author-notes" rid="fn001"><sup>&#x02020;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1224554/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>He</surname> <given-names>Xiang</given-names></name>
<xref ref-type="author-notes" rid="fn001"><sup>&#x02020;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2130555/overview"/>
</contrib> 
<contrib contrib-type="author" corresp="yes">
<name><surname>Yue</surname> <given-names>Zhenzhu</given-names></name>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1215406/overview"/>
</contrib>
</contrib-group>
<aff><institution>Department of Psychology, Sun Yat-sen University</institution>, <addr-line>Guangzhou</addr-line>, <country>China</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Jochen Kaiser, Goethe University Frankfurt, Germany</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Julian Keil, University of Kiel, Germany; Valerio Santangelo, University of Perugia, Italy</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Zhenzhu Yue <email>yuezhzh&#x00040;mail.sysu.edu.cn</email></corresp>
<fn fn-type="other" id="fn001"><p><sup>&#x02020;</sup>These authors have contributed equally to this work and share first authorship</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>23</day>
<month>02</month>
<year>2023</year>
</pub-date>
<pub-date pub-type="collection">
<year>2023</year>
</pub-date>
<volume>17</volume>
<elocation-id>1120668</elocation-id>
<history>
<date date-type="received">
<day>10</day>
<month>12</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>30</day>
<month>01</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2023 Yuan, He and Yue.</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Yuan, He and Yue</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract><p>The interplay between different modalities can help to perceive stimuli more effectively. However, very few studies have focused on how multisensory distractors affect task performance. By adopting behavioral and event-related potentials (ERPs) techniques, the present study examined whether multisensory audiovisual distractors could attract attention more effectively than unisensory distractors. Moreover, we explored whether such a process was modulated by working memory load. Across three experiments, n-back tasks (1-back and 2-back) were adopted with peripheral auditory, visual, or audiovisual distractors. Visual and auditory distractors were white discs and pure tones (Experiments 1 and 2), pictures and sounds of animals (Experiment 3), respectively. Behavioral results in Experiment 1 showed a significant interference effect under high working memory load but not under low load condition. The responses to central letters with audiovisual distractors were significantly slower than those to letters without distractors, while no significant difference was found between unisensory distractor and without distractor conditions. Similarly, ERP results in Experiments 2 and 3 showed that there existed an integration only under high load condition. That is, an early integration for simple audiovisual distractors (240&#x02013;340 ms) and a late integration for complex audiovisual distractors (440&#x02013;600 ms). These findings suggest that multisensory distractors can be integrated and effectively attract attention away from the main task, i.e., interference effect. Moreover, this effect is pronounced only under high working memory load condition.</p></abstract>
<kwd-group>
<kwd>multisensory integration</kwd>
<kwd>audiovisual</kwd>
<kwd>distractor</kwd>
<kwd>working memory load</kwd>
<kwd>n-back task</kwd>
</kwd-group>
<contract-sponsor id="cn001">Fundamental Research Funds for the Central Universities<named-content content-type="fundref-id">10.13039/501100012226</named-content></contract-sponsor>
<counts>
<fig-count count="5"/>
<table-count count="1"/>
<equation-count count="26"/>
<ref-count count="51"/>
<page-count count="13"/>
<word-count count="10782"/>
</counts>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>In our daily lives, we are surrounded by information from different senses, such as audition, vision, touch, and so on. Due to the limited capacity of attention, we cannot process all the information effectively. Previous studies have demonstrated that multisensory stimuli can be integrated and capture attention more effectively than unisensory stimuli (Santangelo and Spence, <xref ref-type="bibr" rid="B41">2007</xref>). For example, multisensory processing shows a clear processing advantage over unisensory processing (ten Oever et al., <xref ref-type="bibr" rid="B51">2016</xref>), yielding more precise representation (Frassinetti et al., <xref ref-type="bibr" rid="B17">2002</xref>), more accurate localization (Van der Stoep et al., <xref ref-type="bibr" rid="B55">2014</xref>), and enhanced stimulus detection (Diederich and Colonius, <xref ref-type="bibr" rid="B13">2004</xref>).</p>
<p>In the past decades, multisensory integration has been extensively investigated by many researchers. However, whether attention is needed for multisensory integration is still a controversial question. Some studies have found that multisensory integration can occur pre-attentively (Caclin et al., <xref ref-type="bibr" rid="B8">2002</xref>; Santangelo and Spence, <xref ref-type="bibr" rid="B41">2007</xref>; Zimmer and Macaluso, <xref ref-type="bibr" rid="B59">2007</xref>; Santangelo et al., <xref ref-type="bibr" rid="B42">2008a</xref>; Van der Burg et al., <xref ref-type="bibr" rid="B52">2008</xref>, <xref ref-type="bibr" rid="B53">2009</xref>; Wahn et al., <xref ref-type="bibr" rid="B56">2017</xref>). For instance, Santangelo and Spence (<xref ref-type="bibr" rid="B41">2007</xref>) and Santangelo et al. (<xref ref-type="bibr" rid="B42">2008a</xref>) investigated attention capture by unisensory or multisensory cues under different perceptual loads. In their spatial cueing task, a visual target was presented after a unisensory (visual or auditory) or multisensory (audiovisual) cue. Participants were required to discriminate the location of the target under low load (no secondary task) or high load conditions (with a concurrent rapid serial visual presentation task). Their results showed that both unisensory and multisensory cues captured attention in the low load condition; whereas only multisensory cues could capture attention in the high load condition. These findings suggest that the multisensory stimuli can be integrated pre-attentively, thus capturing attention more effectively than unisensory stimuli, especially under high perceptual load condition. Similarly, Van der Burg et al. (<xref ref-type="bibr" rid="B52">2008</xref>) adopted a visual search task to examine the influence of perceptual load on multisensory integration. They found that under both low and high perceptual load conditions, the target &#x0201C;popped out&#x0201D; when the visual target was presented concurrently with an auditory pip, suggesting that visual and auditory stimuli were integrated and captured attention effectively irrespective of perceptual load.</p>
<p>However, other studies have shown that multisensory integration can be modulated by attention (Alsius et al., <xref ref-type="bibr" rid="B3">2005</xref>, <xref ref-type="bibr" rid="B2">2014</xref>; Talsma and Woldorff, <xref ref-type="bibr" rid="B47">2005</xref>; Talsma et al., <xref ref-type="bibr" rid="B48">2007</xref>; Hyun et al., <xref ref-type="bibr" rid="B20">2009</xref>; Gibney et al., <xref ref-type="bibr" rid="B19">2017</xref>; Lunn et al., <xref ref-type="bibr" rid="B30">2019</xref>). For instance, Lunn et al. (<xref ref-type="bibr" rid="B30">2019</xref>) investigated the modulation of perceptual load on attention capture by multisensory stimuli. Participants were required to search for the visual targets in the central display or to indicate the location of the visual or audiovisual targets in the periphery display. The visual search display consisted of one target letter and other different letters in the high load condition, whereas one target letter and small placeholder &#x0201C;O&#x0201D;s were included in the display in the low load condition. The multisensory integration was observed in the high load condition but not in the low load condition, indicating that multisensory integration was modulated by attention resources. Similarly, Talsma and Woldorff (<xref ref-type="bibr" rid="B47">2005</xref>) required participants to attend to one side (left or right) and detect oddball targets on that side. Event-related potentials (ERPs) results showed an early audiovisual integration around 100 ms for the attended side but not for the unattended side. They also found integration effects at the time window of 160&#x02013;200 ms and 320&#x02013;420 ms. Moreover, these integration effects were stronger for attended stimuli than for unattended stimuli. These findings suggest that attention modulates the integration of audiovisual stimuli in multiple stages.</p>
<p>One approach to solve the above debate about the role of attention in multisensory integration is to manipulate the attention resources. Working memory task is one of these tasks for modulating the available attention resources (Zimmer and Macaluso, <xref ref-type="bibr" rid="B59">2007</xref>; Michail and Keil, <xref ref-type="bibr" rid="B35">2018</xref>). Previous studies have testified there exists a close relationship between attention and working memory (Downing, <xref ref-type="bibr" rid="B14">2000</xref>; Botta et al., <xref ref-type="bibr" rid="B5">2010</xref>; Brunetti et al., <xref ref-type="bibr" rid="B7">2017</xref>; Oberauer, <xref ref-type="bibr" rid="B38">2019</xref>). Moreover, Santangelo et al. (<xref ref-type="bibr" rid="B43">2006</xref>, <xref ref-type="bibr" rid="B44">2008b</xref>) did not find a more pronounced exogenous orienting effect by multisensory cues than unimodal cues, suggesting that there exists a supramodal spatial attention module that allocates attentional resources towards stimuli from different senses. By using an n-back task as a secondary task, Michail and Keil (<xref ref-type="bibr" rid="B35">2018</xref>) found that the integration of non-speech, audiovisual stimuli was enhanced under reduced attentional resources (high WM load condition), suggesting that top-down attentional control plays an essential role in multisensory integration.</p>
<p>Previous studies have demonstrated that the presence of multisensory stimuli could affect working memory. For example, Botta et al. (<xref ref-type="bibr" rid="B6">2011</xref>) found that spatially congruent multisensory cues showed a more pronounced attentional effect on working memory as compared to unimodal visual cues. This multisensory advantage remained when multisensory stimuli were used as targets for memorizing (Mastroberardino et al., <xref ref-type="bibr" rid="B32">2008</xref>). These results suggest that multisensory integration can facilitate working memory performance. However, few studies have focused on how working memory affects multisensory integration. Thus, in the present study, to investigate the relationship between attention and multisensory processing, we adopted an n-back working memory task to manipulate the attention resources and explored the multisensory processing under different working memory load conditions.</p>
<p>To date, most studies focused on the multisensory integration of task-relevant stimuli, i.e., multisensory stimuli were used as targets. However, multisensory targets are supposed to capture attention because participants voluntarily allocate attention to them to complete the task. It remains unclear whether multisensory distractors can attract attention more effectively than unisensory distractors and whether attention is needed for the multisensory integration of distractors. Although the multisensory integration of targets has been widely studied in recent two decades, only recently has the multisensory integration of distractors been studied. By using a modified multisensory flanker task (for a review, see Merz et al., <xref ref-type="bibr" rid="B33">2021</xref>), Jensen et al. (<xref ref-type="bibr" rid="B21">2019</xref>) and Merz et al. (<xref ref-type="bibr" rid="B34">2019</xref>) found that multisensory integration of task-irrelevant stimuli was modulated by overt attention. Specifically, audiovisual and visuotactile distractors were integrated only when they were presented inside the focus of overt attention. In their follow-up study (Jensen et al., <xref ref-type="bibr" rid="B22">2020</xref>), they found that the audiovisual distractors matching the attentional set induced a significant interference effect. By contrast, this interference effect disappeared when the audiovisual distractors did not match the attentional set, indicating that attention was a key factor in the integration of multisensory distractors. Similarly, by using a central visual search task with peripheral distractors, Lunn et al. (<xref ref-type="bibr" rid="B30">2019</xref>) did not find significant differences in the interference effects between unisensory (visual) and multisensory (audiovisual) distractors. These findings suggest that multisensory integration does not occur when stimuli are task-irrelevant or not attended to. However, by recording the ERPs, Van der Burg et al. (<xref ref-type="bibr" rid="B54">2011</xref>) found an early integration (around 50 ms) of audiovisual distractors. Although the behavioral costs of audiovisual distractors were not significant, this result demonstrated that audiovisual distractors could also be integrated automatically.</p>
<p>The present study aimed to investigate whether multisensory distractors could be integrated and affect attention more effectively than unisensory distractors. Moreover, we explored whether the multisensory integration of audiovisual distractors was modulated by working memory load. In three experiments, 1-back (low load) and 2-back (high load) tasks were adopted. Participants were required to perform the central n-back task while ignoring the peripheral auditory, visual or audiovisual distractors. In Experiment 1, simple white discs and pure tones were used as distractors. To elucidate the neural correlates of the processing of multisensory distractors, Experiment 2 recorded ERPs on the basis of Experiment 1. Given that most of the real-world stimuli contain semantic information and are more complicated compared with simple stimuli, to which extent the results of the simple stimuli can be extended to complex stimuli need to be considered. By using ecological real-life stimuli, previous studies have found that semantic congruence can affect multisensory processing, suggesting the necessity of using semantic real-word stimuli (Mastroberardino et al., <xref ref-type="bibr" rid="B31">2015</xref>; Kvasova et al., <xref ref-type="bibr" rid="B25">2019</xref>; Almadori et al., <xref ref-type="bibr" rid="B1">2021</xref>). Therefore, to improve the ecology of the stimuli, another ERP experiment (Experiment 3) was conducted by adopting pictures of animals and the sounds they made as distractors. We hypothesized that compared with unisensory distractors, multisensory audiovisual distractors were more effective to attract attention and were prone to interfere with the performance of the n-back task. Moreover, the working memory load was expected to modulate this interference effect. That is, the interference effect should be stronger under high load conditions, especially for audiovisual distractors. For the ERPs, the integration effect should be more pronounced in the high load conditions than in the low load conditions. Moreover, the integration of audiovisual distractors should be modulated by the complexity of distractor stimuli, as reflected by the time window of the significant integration effect. That is, the time window of the significant integration effect should be observed later for complex distractors than for simple distractors.</p>
</sec>
<sec id="s2">
<title>2 Experiment 1</title>
<sec id="s2-1">
<title>2.1 Method</title>
<sec id="s2-1-1">
<title>2.1.1 Participants</title>
<p>According to the effect size of a similar study [<inline-formula><mml:math id="M1"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.10; Experiment 4 in Lunn et al. (<xref ref-type="bibr" rid="B30">2019</xref>)], a sample size estimation was done using G*power software (Faul et al., <xref ref-type="bibr" rid="B16">2009</xref>). The result revealed that a sample of 26 participants was required to at least detect an interaction with an effect size of <inline-formula><mml:math id="M2"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.10 (<italic>&#x003B1;</italic> = 0.05, 1-<italic>&#x003B2;</italic> = 0.80). Twenty-nine healthy college students participated in the experiment. They reported a normal or corrected-to-normal vision and normal hearing. Three participants were excluded because the accuracy rates were lower than 70%. Data of 26 participants (19 females; mean age = 19.92 years, SD = 1.74, range = 18&#x02013;24 years) entered the final analysis. All participants signed informed consent and were paid 25 RMB. The study was approved by the Ethics Committee of the Department of Psychology, Sun Yat-sen University.</p>
</sec>
<sec id="s2-1-2">
<title>2.1.2 Apparatus and stimuli</title>
<p>The experiment was controlled by E-prime 2.0 software<xref ref-type="fn" rid="fn0001"><sup>1</sup></xref>. Participants sat 60 cm in front of a 23-inch monitor (1,920 &#x000D7; 1,080; 60 Hz) in a sound-attenuated, dimly lit room. The auditory stimuli were generated by Adobe Audition CC 2019 software, sampled at 44.1 kHz, and quantized to 16 bits. Before the experiment, the sound was tuned to a comfortable volume for all participants (range: 35&#x02013;45 dB).</p>
<p>All consonants except &#x0201C;Y&#x0201D; were used (1&#x000B0; &#x000D7; 1.4&#x000B0;of visual angle) in the n-back task. Auditory distractors (1,000 Hz pure tone) were presented for 200 ms either on the left or right side equiprobably <italic>via</italic> headphones (SONY MDR-XB450). Visual distractors were white discs (diameter: 1.9&#x000B0; visual angle), presented at an eccentricity of 7.5&#x000B0; degrees (screen center to discs center). All visual distractors were presented either to the left or right side of the central letters with equal probability for 200 ms. In the audiovisual distractors condition, auditory and visual distractors were presented at the same side concurrently.</p>
</sec>
<sec id="s2-1-3">
<title>2.1.3 Design and procedure</title>
<p>A 2 (Load: low vs. high) &#x000D7; 4 (Distractor type: auditory, visual, audiovisual vs. no distractor) within-participants design was adopted. The 1-back and 2-back tasks were used to manipulate the working memory load. The trial scheme is shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. Each letter series started with a cross fixation presented at the center of the screen for 500 ms. The fixation was then replaced by a stream of fourteen letters. Each letter was presented for 200 ms with an inter-stimulus interval (ISI) of 1,800 ms. Participants were required to memorize the stream of the letter and to report whether the current target letter was the same or not as the letter presented one or two steps back, in 1-back and 2-back tasks respectively. They were instructed to respond by pressing one of two buttons (LB and RB) with the joystick as quickly and accurately as possible. Response keys on the joystick were counterbalanced between participants.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p><bold>(A)</bold> Trial scheme in Experiment 1 and Experiment 2 (2-back task in Experiment 1 as an example). <bold>(B)</bold> Stimuli used in Experiment 3.</p></caption>
<graphic xlink:href="fnint-17-1120668-g0001.tif"/>
</fig>
<p>In addition to the central visual targets, an auditory, visual, or audiovisual distractor was also presented concurrently with the letter in 75% of all trials. They were presented equiprobably at the left or right side of the letter for 200 ms. Participants were required to ignore the peripheral distractor and concentrate on the central letter. In the remaining 25% of all trials, no distractors were presented. Overall, the experiment consisted of 8 blocks with 8 letter series each. Participants had practiced before the formal experiments.</p>
</sec>
<sec id="s2-1-4">
<title>2.1.4 Data analysis</title>
<p>Reaction times (RTs) and accuracy rate (ACC) were calculated separately for each experimental condition. For all participants, RTs of correct responses between 100 and 1,800 ms were included in the analysis. Besides, RTs exceeding &#x000B1; 3 SD of each participant&#x02019;s mean reaction time in each experimental condition were removed. Analyses of variance (ANOVAs) were calculated separately for mean RTs and mean ACC with factors of Load (low vs. high) and Distractor type (auditory, visual, audiovisual vs. no distractor). The interference effect was calculated by subtracting the mean RT without distractors from the mean RT with distractors. If necessary, the Greenhouse-Geisser method was adopted to correct degrees of freedom. Besides, Bonferroni corrections were used for <italic>post-hoc</italic> pair-wise comparisons and simple effects.</p>
</sec>
</sec>
<sec id="s2-2">
<title>2.2 Results</title>
<p>For RTs, the ANOVA revealed a significant main effect of Load [<italic>F</italic><sub>(1, 25)</sub> = 155.46, <italic>p</italic> &#x0003C; 0.001, <inline-formula><mml:math id="M3"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.86], suggesting that responses in the low load condition were faster than those in the high load condition (<italic>M</italic> = 574.05 vs. 773.49 ms, SE = 18.01 vs. 29.78). We also found a significant interaction between Load and Distractor type [<italic>F</italic><sub>(3, 75)</sub> = 3.51, <italic>p</italic> &#x0003C; 0.05, <inline-formula><mml:math id="M4"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.12; <xref ref-type="fig" rid="F2">Figure 2</xref>]. Follow-up analyses showed that responses to letters with audiovisual distractors were significantly slower than to letters without distractors in the high load condition (<italic>M</italic> = 781.44 vs. 762.94 ms, SE = 30.20 vs. 29.32, <italic>t</italic><sub>(25)</sub> = 3.05, <italic>p</italic> &#x0003C; 0.05); whereas in the low load condition, no significant differences were found between these two conditions. No other significant <italic>post-hoc</italic> pair-wise comparisons or main effect [Distractor type: <italic>F</italic><sub>(2.36, 59.08)</sub> = 1.98, <italic>p</italic> = 0.14] was found.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Means and standard errors for the interference effect (RTs with distractors minus RTs without distractors) under auditory, visual, and audiovisual distractors conditions. The interference effect was shown under low and high working memory load, respectively. * indicates <italic>p</italic> &#x0003C; 0.05. RTs, reaction times.</p></caption>
<graphic xlink:href="fnint-17-1120668-g0002.tif"/>
</fig>
<p>For ACC, the ANOVA revealed a significant main effect of Load [<italic>F</italic><sub>(1, 25)</sub> = 34.59, <italic>p</italic> &#x0003C; 0.001, <inline-formula><mml:math id="M5"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.58], reflecting that higher ACC was observed in the low load condition than in the high load condition (<italic>M</italic> = 0.95 vs. 0.92, SE = 0.01 vs. 0.01). Neither the main effect of Distractor type [<italic>F</italic><sub>(2.38, 59.54)</sub> = 1.55, <italic>p</italic> = 0.22, <inline-formula><mml:math id="M6"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.06] nor the interaction between Load and Distractor type [<italic>F</italic><sub>(3, 75)</sub> = 1.57, <italic>p</italic> = 0.20, <inline-formula><mml:math id="M7"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.06] was significant.</p>
</sec>
<sec id="s2-3">
<title>2.3 Discussion</title>
<p>Overall, we found a significant interference effect of multisensory audiovisual distractors in Experiment 1, as reflected by longer RTs in the working memory task when audiovisual distractors were presented compared to no distractors were presented. However, such a significant interference effect was shown only under high load conditions rather than under low load conditions, suggesting that AV distractors easily draw attention away from the main task under high load conditions compared to the absence of distractors.</p>
<p>It should be noted that a significant response difference between unisensory and multisensory signals does not necessarily mean that the multisensory stimuli have been integrated. According to previous reviews (Stein et al., <xref ref-type="bibr" rid="B45">2010</xref>; Keil and Senkowski, <xref ref-type="bibr" rid="B23">2018</xref>), multisensory integration is defined more strictly as the neural process difference between the unisensory and multisensory stimuli. Thus, to further investigate the neural correlate of the effect of audiovisual distractors, that is, whether the multisensory integration happens or not, we recorded ERPs in Experiment 2.</p>
</sec>
</sec>
<sec id="s3">
<title>3 Experiment 2</title>
<sec id="s3-1">
<title>3.1 Method</title>
<sec id="s3-1-1">
<title>3.1.1 Participants</title>
<p>Based on the effect size of a similar study [<inline-formula><mml:math id="M8"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.10; Experiment 4 in Lunn et al. (<xref ref-type="bibr" rid="B30">2019</xref>)], a sample size estimation was done using G*power software (Faul et al., <xref ref-type="bibr" rid="B16">2009</xref>). The result revealed that a sample of 26 participants was required to at least detect an interaction with an effect size of <inline-formula><mml:math id="M9"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.10 (<italic>&#x003B1;</italic> = 0.05, 1-<italic>&#x003B2;</italic> = 0.80). A new group of thirty-two healthy college students participated in the experiment. They had a normal or corrected-to-normal vision and normal hearing. Three participants were excluded because of excessive (>25%) EEG artifacts. Data of 29 participants (16 females; mean age = 20.62 years, SD = 1.99, range = 18&#x02013;26 years) entered the final analysis. All participants signed informed consent and were paid 75 RMB. The study was approved by the Ethics Committee of the Department of Psychology, Sun Yat-sen University.</p>
</sec>
<sec id="s3-1-2">
<title>3.1.2 Apparatus and stimuli</title>
<p>The experimental apparatus and stimuli were the same as those in Experiment 1 except for the eccentricity of the distractors (4&#x000B0;), and the presenting mode of the auditory stimuli. Specifically, auditory distractors were presented at either the left or right side equiprobably <italic>via</italic> two invisible loudspeakers (Creative inspire T12), which were placed at the source location of the visual distractors behind the screen. Before the experiment, the sound was tuned to a comfortable volume for all participants (range: 65&#x02013;75 dB).</p>
</sec>
<sec id="s3-1-3">
<title>3.1.3 Design and procedure</title>
<p>A 2 (Load: low vs. high) &#x000D7; 4 (Distractor type: auditory, visual, audiovisual vs. no distractor) within-participants design was adopted. There were 105 trials for each experimental condition. The procedure and task were the same as those in Experiment 1 (<xref ref-type="fig" rid="F1">Figure 1</xref>).</p>
</sec>
<sec id="s3-1-4">
<title>3.1.4 Electroencephalogram (EEG) recording and preprocessing</title>
<p>The EEG was recorded from 64 Ag-AgCl electrodes mounted in an elastic cap (Easy Cap, Germany) with a NeuroScan SynAmps2 Amplifier (Scan 4.5, Neurosoft Labs, Inc. Virginia, USA). A left earlobe electrode was used as an online reference. The ground electrode was located on the forehead. Vertical eye movements were monitored with two electrodes upper and below the right eye. Horizontal eye movements were recorded with two electrodes placed at the outer canthi of each eye. Electrode impedance was kept below 5 k&#x003A9; for all electrodes. Online recordings were bandpass filtered at 0.05&#x02013;100 Hz (12 dB/oct, 40 dB/dec) and sampled at 500 Hz. During the experiment, participants were instructed to fixate on the center of the monitor and try not to make horizontal or vertical eye movements.</p>
<p>The offline analysis of EEG data was performed using Matlab R2016b and eeglab 14.1.2b<xref ref-type="fn" rid="fn0002"><sup>2</sup></xref>. First, all scalp electrodes were re-referenced to the average of left and right earlobes. Then, the continuous EEG was bandpass filtered (IIR Butterworth, filter order = 2) at 0.05&#x02013;30 Hz. An infomax independent component analysis (ICA) algorithm (Bell and Sejnowski, <xref ref-type="bibr" rid="B4">1995</xref>) was applied for correcting eye movement artifacts. The SASICA plugin with ADJUST was used to identify the artifact component. Furthermore, the interval of 0&#x02013;200 ms prior to the distractors served as the baseline; EEG signal epochs ended 800 ms after the onset of the distractor stimuli, yielding a total epoch of 1 s. Finally, trials with voltages exceeding &#x000B1; 100 &#x003BC;V were excluded from ERP averages. The remaining epochs to eight different conditions were averaged separately for each participant with baseline corrections. In the present experiment, the average artifact rejection rate was 3.23% of all trials (SD = 5.02, range = 0&#x02013;18.2%).</p>
</sec>
<sec id="s3-1-5">
<title>3.1.5 Data analysis</title>
<p>Behavioral data analyses were identical to Experiment 1. Analyses of variance (ANOVAs) were calculated separately for mean RTs and mean ACC with within-participants factors of Load (low vs. high) and Distractor type (auditory, visual, audiovisual vs. no distractor).</p>
<p>For the ERPs, to control for the overlap and generic cognitive process (such as contingent negative variation, CNV), the ERPs elicited by no distractor trials were subtracted from the ERPs elicited by auditory (A), visual (V), and audiovisual (AV) distractors, respectively. Then, to estimate the multisensory integration effect, the ERPs elicited by A distractors and V distractors were summated (A + V) and compared with the ERPs elicited by AV distractors. Specifically, the audiovisual distractors were integrated if significant differences were found between the (A + V) ERPs and the AV ERPs (Giard and Peronnet, <xref ref-type="bibr" rid="B18">1999</xref>; Stevenson et al., <xref ref-type="bibr" rid="B46">2014</xref>).</p>
<p>Time windows and electrodes were selected based on the previous studies (Talsma and Woldorff, <xref ref-type="bibr" rid="B47">2005</xref>; Van der Burg et al., <xref ref-type="bibr" rid="B54">2011</xref>), the grand average ERPs and the topographic map. Previous studies have found three phases of effects of integration and/or attention beginning at around 160 ms, and peaking at 190 ms (scalp positivity), 250 ms (negativity), and 300&#x02013;500 ms (positivity) after stimulus onset (Talsma and Woldorff, <xref ref-type="bibr" rid="B47">2005</xref>). We also did a mass-univariate statistical analysis with corrections based on previous studies. Specifically, we did ANOVAs with factors of Load and Distractor type at each electrode and each time-point across participants. To avoid the type-I error due to the large number of tests, the multisensory integration effects were thought to be significant only when the <italic>p</italic>-value was smaller than 0.05 at 10 (&#x0007E;20 ms) or more continuous time points on at least two nearby electrodes (Van der Burg et al., <xref ref-type="bibr" rid="B54">2011</xref>; Alsius et al., <xref ref-type="bibr" rid="B2">2014</xref>). Two time windows of 240&#x02013;340 ms (electrode FP<sub>Z</sub>) and 450&#x02013;600 ms (electrode F<sub>Z</sub>) were selected. To further test the hypothesis whether multisensory integration should be more pronounced in the high load condition than in the low load condition, the mean amplitudes in these time windows were analyzed by ANOVAs with factors of Load (low vs. high) and Distractor type (A + V vs. AV). If necessary, the Greenhouse-Geisser method was adopted to correct degrees of freedom. Besides, Bonferroni corrections were used for <italic>post-hoc</italic> pair-wise comparisons and simple effects.</p>
</sec>
</sec>
<sec id="s3-2">
<title>3.2 Results</title>
<sec id="s3-2-1">
<title>3.2.1 Behavioral performance</title>
<p>The overall results of ANOVA are shown in <xref ref-type="table" rid="T1">Table 1</xref>. For RTs, ANOVA revealed a significant main effect of Load [<italic>F</italic><sub>(1, 28)</sub> = 52.86, <italic>p</italic> &#x0003C; 0.001, <inline-formula><mml:math id="M10"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.65], indicating slower responses in high load condition than in low load condition (<italic>M</italic> = 790.25 vs. 636.85 ms, SE = 31.97 vs. 22.90). Neither the main effect of Distractor type [<italic>F</italic><sub>(2.44, 68.37)</sub> = 0.68, <italic>p</italic> = 0.54] nor the interaction between Load and Distractor type [<italic>F</italic><sub>(3, 84)</sub> = 0.30, <italic>p</italic> = 0.83] was significant.</p>
<table-wrap id="T1" position="float">
<label>Table 1</label>
<caption><p>Results of the ANOVA in Experiments 2 and 3.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left"></th>
<th align="left" colspan="2">Behavioral results</th>
<th align="left" colspan="2">ERP results</th>
</tr>
<tr>
<th align="left"></th>
<th align="left">RT</th>
<th align="left">ACC</th>
<th align="left">Early integration</th>
<th align="left">Late integration</th>
</tr>
<tr>
<th align="left" colspan="5">Experiment 2</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">Load</td>
<td align="center">52.86 (&#x0003C;0.001)</td>
<td align="center">17.81 (&#x0003C;0.001)</td>
<td align="center">1.21 (0.28)</td>
<td align="center">3.89 (0.06)</td>
</tr>
<tr>
<td align="left">Distractor type</td>
<td align="center">0.68 (0.57)</td>
<td align="center">3.68 (0.02)</td>
<td align="center">0.24 (0.63)</td>
<td align="center">2.19 (0.15)</td>
</tr>
<tr>
<td align="left">Load &#x000D7; Distractor type</td>
<td align="center">0.30 (0.83)</td>
<td align="center">0.75 (0.52)</td>
<td align="center">4.57 (0.04)</td>
<td align="center">0.29 (0.59)</td>
</tr>
<tr>
<td align="center" colspan="5">Experiment 3</td>
</tr>
<tr>
<td align="left">Load</td>
<td align="center">115.59 (&#x0003C;0.001)</td>
<td align="center">18.88 (&#x0003C;0.001)</td>
<td align="center">0.64 (0.50)</td>
<td align="center">2.13 (0.16)</td>
</tr>
<tr>
<td align="left">Distractor type</td>
<td align="center">8.20 (&#x0003C;0.001)</td>
<td align="center">0.57 (0.64)</td>
<td align="center">0.54 (0.47)</td>
<td align="center">2.99 (0.09)</td>
</tr>
<tr>
<td align="left">Load &#x000D7; Distractor type</td>
<td align="center">0.24 (0.87)</td>
<td align="center">2.04 (0.11)</td>
<td align="center">0.26 (0.61)</td>
<td align="center">4.69 (0.04)</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>F</italic>-value and <italic>p</italic>-value (in parentheses).</p>
</table-wrap-foot>
</table-wrap>
<p>For ACC, the main effect of Load was significant [<italic>F</italic><sub>(1, 28)</sub> = 17.81, <italic>p</italic> &#x0003C; 0.001, <inline-formula><mml:math id="M11"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.39], indicating ACC in low load condition was higher than in high load condition (<italic>M</italic> = 0.94 vs. 0.90, SEs = 0.01). The main effect of Distractor type was significant [<italic>F</italic><sub>(3, 84)</sub> = 3.68, <italic>p</italic> &#x0003C; 0.01, <inline-formula><mml:math id="M12"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.12]. However, <italic>post-hoc</italic> analyses did not reveal any significant pair-wise comparisons. The interaction between Load and Distractor [<italic>F</italic><sub>(3, 84)</sub> = 0.75, <italic>p</italic> = 0.52] type was not significant, either.</p>
</sec>
<sec id="s3-2-2">
<title>3.2.2 ERP results: overlap correction</title>
<p>To control for the differential overlap and generic cognitive process, overlap correction was done before summating the A and V ERPs (Giard and Peronnet, <xref ref-type="bibr" rid="B18">1999</xref>; Talsma and Woldorff, <xref ref-type="bibr" rid="B47">2005</xref>; Van der Burg et al., <xref ref-type="bibr" rid="B54">2011</xref>; Stevenson et al., <xref ref-type="bibr" rid="B46">2014</xref>). To show the necessity of this overlap correction, ERPs elicited by stimuli with A, V, and AV distractors under high load conditions at the C<sub>Z</sub> electrode were averaged, and then ERPs elicited by stimuli without distractor under high load conditions at the C<sub>Z</sub> electrode were subtracted from the averaged ERP waveform. The effect of overlap correction was shown in <xref ref-type="fig" rid="F3">Figure 3</xref>.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Overlap correction. Averaged ERPs with distractors (averaged with overlaps), ERPs without distractors, and corrected averaged ERPs (averaged ERPs with distractors minus ERPs without distractors) under high load conditions at the C<sub>Z</sub> electrode. ERPs, event-related potentials.</p></caption>
<graphic xlink:href="fnint-17-1120668-g0003.tif"/>
</fig>
<p>To further quantify the effectiveness of overlap correction, we tested whether the ERPs elicited by stimuli with (A + V) and AV distractors differed before the onset of the distractor. Theoretically, no significant differences in the mean amplitudes should be found during any time window of &#x02212;200&#x02013;0 ms. Besides, the averaged ERPs of distractors should not differ from 0 &#x003BC;V. Thus, the time window of &#x02212;20&#x02013;0 ms at the C<sub>Z</sub> electrode was selected. The one-sample <italic>t</italic>-test was used to test whether the averaged ERPs elicited by stimuli with distractors differed from 0 &#x003BC;V. ANOVA with factors of Load (low vs. high) and Distractor type (A + V vs. AV) was used to test whether the ERPs elicited by stimuli with (A + V) and AV distractors differed from each other. When no-distractor ERPs were not subtracted from ERPs of (A + V) and AV distractors, although no significant results were found for the ANOVA (main effect of Distractor type: [<italic>F</italic><sub>(1, 28)</sub> = 1.56, <italic>p</italic> = 0.22, <inline-formula><mml:math id="M13"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> =0.05]), the one-sample <italic>t</italic>-test showed that the averaged ERPs with distractors differed significantly from 0 &#x003BC;V [<italic>t</italic><sub>(28)</sub> = &#x02212;5.80, <italic>p</italic> &#x0003C; 0.001; <italic>M</italic> = &#x02212;0.49 &#x003BC;V, SE = 0.08]. However, after subtracting the no-distractor ERPs from the (A + V) and AV ERPs, neither the <italic>t</italic>-test [<italic>t</italic><sub>(28)</sub> = &#x02212;1.77, <italic>p</italic> = 0.09] nor the main effect of Distractor type [<italic>F</italic><sub>(1, 28)</sub> = 0.06, <italic>p</italic> = 0.81, <inline-formula><mml:math id="M14"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.03] was significant. These results showed that subtracting the ERPs elicited by no-distractor trials from the ERPs elicited by A, V, and AV distractor trials could effectively remove the overlap due to the generic cognitive process.</p>
</sec>
<sec id="s3-2-3">
<title>3.2.3 ERP results: early integration</title>
<p>The results of the ANOVA conducted at each time epoch are reported in <xref ref-type="table" rid="T1">Table 1</xref>. The ANOVA of mean amplitudes for the time window of 240&#x02013;340 ms showed a significant interaction between Load and Distractor type (<italic>F</italic><sub>(1, 28)</sub> = 4.57, <italic>p</italic> &#x0003C; 0.05, <inline-formula><mml:math id="M15"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.14; see <xref ref-type="fig" rid="F4">Figure 4</xref>). Follow-up analyses showed that the mean amplitudes of (A + V) ERPs were more positive than those in the AV condition while the working memory load was high (<italic>M</italic> = 0.85 vs. 0.08 &#x003BC;V, SE = 0.60 vs. 0.41; <italic>t</italic><sub>(28)</sub> = 2.10, <italic>p</italic> &#x0003C; 0.05). However, under low load conditions, no significant differences in mean amplitudes were observed between these two conditions. These results showed that audiovisual distractors were integrated under high load conditions but not under low load conditions, suggesting that the working memory load modulated the integration of audiovisual distractors. Neither the main effect of Load [<italic>F</italic><sub>(1, 28)</sub> = 1.21, <italic>p</italic> > 0.05] nor the main effect of Distractor type [<italic>F</italic><sub>(1, 28)</sub> = 0.24, <italic>p</italic> > 0.05] was significant.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p><bold>(A)</bold> Grand-average distractor-synchronized ERP waveforms at each experimental condition. Time window of 240&#x02013;340 ms is highlighted in the box (electrode FP<sub>Z</sub>). Time zero on the x-axis corresponds to distractor stimuli onset. <bold>(B)</bold> Mean amplitudes and standard errors for different conditions during time window of 240&#x02013;340 ms (electrode FP<sub>Z</sub>). * indicates <italic>p</italic> &#x0003C; 0.05.</p></caption>
<graphic xlink:href="fnint-17-1120668-g0004.tif"/>
</fig>
<p>In addition, the ANOVAs of mean amplitudes for the late time window 450&#x02013;600 ms showed no significant interaction between Load and Distractor type [<italic>F</italic><sub>(1, 28)</sub> = 0.29, <italic>p</italic> > 0.05]. Neither the main effect of Load [<italic>F</italic><sub>(1, 28)</sub> = 3.89, <italic>p</italic> > 0.05] nor the main effect of Distractor type [<italic>F</italic><sub>(1, 28)</sub> = 2.19, <italic>p</italic> > 0.05] was significant.</p>
</sec>
</sec>
<sec id="s3-3">
<title>3.3 Discussion</title>
<p>By adopting the ERP technique, we investigated whether audiovisual distractors could be integrated and how working memory load affected the multisensory integration. Although no significant interactions were found in behavioral performances, ERP results showed that working memory load modulated the integration of audiovisual distractors, that is, the audiovisual distractors could be integrated only under high load conditions. Specifically, under high load conditions, after the overlap correction, significant differences in the ERP mean amplitudes were found between the (A + V) distractor conditions and the AV distractor conditions. This finding indicated that audiovisual distractors were integrated under high working memory load conditions, while such a pattern was eliminated under low load condition. The results in Experiment 2 are in line with the findings in Experiment 1, that is, the capacity of inhibiting distractors is reduced under high load condition.</p>
<p>Experiment 2 demonstrated that the integration of simple audiovisual distractors happened at the time window of 240&#x02013;340 ms after the onset of the distractors, indicating an early integration of audiovisual distractors under high load conditions. This time window of audiovisual integration is consistent with previous studies focusing on the integration of audiovisual targets (Teder-S&#x000E4;lej&#x000E4;rvi et al., <xref ref-type="bibr" rid="B50">2002</xref>; Talsma and Woldorff, <xref ref-type="bibr" rid="B47">2005</xref>; Van der Burg et al., <xref ref-type="bibr" rid="B54">2011</xref>). For instance, Teder-S&#x000E4;lej&#x000E4;rvi adopted a multisensory oddball paradigm and found integration at the time window of 242&#x02013;226 ms and 300&#x02013;400 ms after the onset of the audiovisual targets.</p>
<p>The distractors used in Experiments 1 and 2 are simple stimuli (visual discs and auditory pure tones). However, most of the stimuli we interact with in real life are complex stimuli, and to which extent the results of the integration of simple stimuli can be extended to complex stimuli should be considered (Koelewijn et al., <xref ref-type="bibr" rid="B24">2010</xref>). Thus, in Experiment 3, by adopting animal pictures and the sounds they made as distractors, we further investigated how the multisensory audiovisual distractor affected attention.</p>
</sec>
</sec>
<sec id="s4">
<title>4 Experiment 3</title>
<sec id="s4-1">
<title>4.1 Method</title>
<sec id="s4-1-1">
<title>4.1.1 Participants</title>
<p>In reference to the effect size of a similar study [<inline-formula><mml:math id="M16"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.10; Experiment 4 in Lunn et al. (<xref ref-type="bibr" rid="B30">2019</xref>)], a sample size estimation was done using G*power software (Faul et al., <xref ref-type="bibr" rid="B16">2009</xref>). The result revealed that a sample of 26 participants was required to at least detect an interaction with an effect size of <inline-formula><mml:math id="M17"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.10 (<italic>&#x003B1;</italic> = 0.05, 1-<italic>&#x003B2;</italic> = 0.80). Another new group of thirty-seven healthy college students participated in the experiment. They had a normal or corrected-to-normal vision and normal hearing. Two participants were excluded because of equipment problems. Another three participants were excluded because of excessive (>25%) EEG artifacts. Data of 32 participants (21 females; mean age = 20.19 years, SD = 2.15, range = 18&#x02013;27 years) entered the final analysis. Participants signed informed consent and were paid 75 RMB. The study was approved by the Ethics Committee of the Department of Psychology, Sun Yat-sen University.</p>
</sec>
<sec id="s4-1-2">
<title>4.1.2 Apparatus and stimuli</title>
<p>The experimental apparatus and stimuli were the same as those in Experiment 2 except for the type of distractors. Specifically, visual distractors consisted of pictures of a cat or a dog. They were presented for 500 ms at an eccentricity of 4&#x000B0; degrees (screen center to image center). Auditory distractors were the sounds that the animals made. All auditory stimuli were presented at either the left or right side equiprobably for 500 ms <italic>via</italic> two invisible loudspeakers (Creative inspire T12) placed at the source location of the visual distractors behind the screen. In the multisensory audiovisual distractors condition, both auditory and visual distractors of the same animal were presented on the same side concurrently. Before the experiment, the sound was tuned to a comfortable volume for all participants (range: 65&#x02013;75 dB).</p>
</sec>
<sec id="s4-1-3">
<title>4.1.3 Design and procedure</title>
<p>A 2 (Load: low vs. high) &#x000D7; 4 (Distractor type: auditory, visual, audiovisual vs. no distractor) within-participants design was adopted. There were 105 trials for each experimental condition. In order to convey semantic information clearly, central letters and peripheral distractors were both presented for 500 ms. The stimuli used as distractors are shown in <xref ref-type="fig" rid="F1">Figure 1B</xref>. The procedure and task were the same as those in Experiment 1.</p>
</sec>
<sec id="s4-1-4">
<title>4.1.4 Electroencephalogram (EEG) recording and preprocessing</title>
<p>EEG recording and preprocessing were identical to Experiment 2. The average artifact rejection rate in the present experiment was 2.27% of all trials (SD = 4.19, range = 0&#x02013;20.4%). ERPs in each experimental condition were averaged separately for each participant.</p>
</sec>
<sec id="s4-1-5">
<title>4.1.5 Data analysis</title>
<p>Behavioral data analyses were identical to Experiment 2. Analyses of variance (ANOVAs) were calculated separately for mean RTs and mean ACC with within-participants factors of Load (low vs. high) and Distractor type (auditory, visual, audiovisual vs. no distractor).</p>
<p>For the ERPs, the data analyses were identical to Experiment 2. After the overlap correction, time windows and electrodes were selected based on the previous studies (Talsma and Woldorff, <xref ref-type="bibr" rid="B47">2005</xref>; Van der Burg et al., <xref ref-type="bibr" rid="B54">2011</xref>), the grand average ERPs and the topographic map. We did a mass-univariate statistical analysis (ANOVAs with factors of Load and Distractor type at each electrode and each time-point across participants) with correction based on previous studies. That is, the multisensory integration effects were thought to be significant only when the <italic>p</italic>-value was smaller than 0.05 at 10 (&#x0007E;20 ms) or more continuous time points on at least two nearby electrodes (Van der Burg et al., <xref ref-type="bibr" rid="B54">2011</xref>; Alsius et al., <xref ref-type="bibr" rid="B2">2014</xref>). Two time windows of 250&#x02013;330 ms (averaged across electrodes C<sub>Z</sub> and FC<sub>Z</sub>) and 440&#x02013;600 ms (averaged across electrodes AF7, F5, and F7) were selected. To further test the hypothesis whether multisensory integration should be more pronounced in the high load condition than in the low load condition, the mean amplitudes in these two time windows were analyzed by ANOVAs with factors of Load (low vs. high) and Distractor type (A + V vs. AV), respectively. If necessary, Greenhouse-Geisser method was adopted to correct degrees of freedom. Besides, Bonferroni corrections were used for <italic>post-hoc</italic> pair-wise comparisons and simple effects.</p>
</sec>
</sec>
<sec id="s4-2">
<title>4.2 Results</title>
<sec id="s4-2-1">
<title>4.2.1 Behavioral performance</title>
<p>The overall results of ANOVA are shown in <xref ref-type="table" rid="T1">Table 1</xref>. For RTs, ANOVA revealed a significant main effect of Load [<italic>F</italic><sub>(1, 31)</sub> = 115.59, <italic>p</italic> &#x0003C; 0.001, <inline-formula><mml:math id="M18"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.79], indicating the responses were slower in the high load condition than in the low load condition (<italic>M</italic> = 700.72 vs. 562.29 ms, SE = 26.62 vs. 17.92). The main effect of Distractor type was significant [<italic>F</italic><sub>(3, 93)</sub> = 8.20, <italic>p</italic> &#x0003C; 0.001, <inline-formula><mml:math id="M19"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.21], indicating the responses to letters with visual distractors (<italic>M</italic> = 639.35 ms, SE = 21.79) were slower than to letters with auditory, audiovisual distractors and with no distractors (auditory: <italic>M</italic> = 625.15 ms, SE = 21.76; audiovisual: <italic>M</italic> = 630.29 ms, SE = 21.80; no distractors: <italic>M</italic> = 631.22 ms, SE = 21.98). The interaction between Load and Distractor Type was not significant [<italic>F</italic><sub>(3, 93)</sub> = 0.24, <italic>p</italic> = 0.87, <inline-formula><mml:math id="M20"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.01].</p>
<p>For ACC, there was a significant main effect of Load [<italic>F</italic><sub>(1, 31)</sub> = 18.88, <italic>p</italic> &#x0003C; 0.001, <inline-formula><mml:math id="M21"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.38], indicating that ACC in the low load condition was higher than in the high load condition (<italic>M</italic> = 0.94 vs. 0.89, SEs = 0.01). Neither the main effect of Distractor type [<italic>F</italic><sub>(3, 93)</sub> = 0.57, <italic>p</italic> = 0.64, <inline-formula><mml:math id="M22"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.02] nor the interaction [<italic>F</italic><sub>(2, 93)</sub> = 2.04, <italic>p</italic> = 0.11, <inline-formula><mml:math id="M23"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.06] was significant.</p>
</sec>
<sec id="s4-2-2">
<title>4.2.2 ERP results: overlap correction</title>
<p>Similar to Experiment 2, ERPs elicited by stimuli with A, V, and AV distractors under high load conditions at the C<sub>Z</sub> electrode were averaged, and then ERPs elicited by stimuli without distractors in the high load condition at the C<sub>Z</sub> electrode were subtracted from the averaged ERP waveform.</p>
<p>To further quantify the effectiveness of overlap correction, we tested whether the ERPs elicited by stimuli with (A + V) and AV distractors differed before the onset of the distractor. The time window of &#x02212;20&#x02013;0 ms at the C<sub>Z</sub> electrode was selected. The one-sample <italic>t</italic>-test was used to test whether the averaged ERPs elicited by stimuli with distractors differed from 0 &#x003BC;V. The two-way ANOVA with factors of Load (low vs. high) and Distractor type (A + V vs. AV) was used to test whether the ERPs elicited by stimuli with (A + V) and AV distractors differed from each other. When no-distractor ERPs were not subtracted from ERPs of (A + V) and AV distractors, the one-sample <italic>t</italic>-test showed that the averaged ERPs with distractors differed significantly from 0 &#x003BC;V [<italic>t</italic><sub>(31)</sub> = &#x02212;5.33, <italic>p</italic> &#x0003C; 0.001; <italic>M</italic> = &#x02212;0.51 &#x003BC;V, SE = 0.10]. The ANOVA also showed a significant main effect of Distractor type [<italic>F</italic><sub>(1, 31)</sub> = 7.12, <italic>p</italic> &#x0003C; 0.05, <inline-formula><mml:math id="M24"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.19], indicating more negative ERPs elicited by stimuli with (A + V) distractors than with AV distractors (<italic>M</italic> = &#x02212;0.95 vs. &#x02212;0.49 &#x003BC;V, SE = 0.22 vs. 0.10; <italic>t</italic><sub>(31)</sub> = 2.66, <italic>p</italic> &#x0003C; 0.05). However, after subtracting the no-distractor ERPs from the (A + V) and AV ERPs, neither the <italic>t</italic>-test [<italic>t</italic><sub>(31)</sub> = 0.12, <italic>p</italic> = 0.90] nor the main effect of Distractor type [<italic>F</italic><sub>(1, 31)</sub> = 0.18, <italic>p</italic> = 0.67, <inline-formula><mml:math id="M25"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.01] was significant. These results showed that subtracting the ERPs elicited by no-distractor trials from the ERPs elicited by A, V, and AV distractor trials effectively removed the overlap due to the generic cognitive process.</p>
</sec>
<sec id="s4-2-3">
<title>4.2.3 ERP results: late integration</title>
<p>The results of the ANOVA conducted at each time epoch are reported in <xref ref-type="table" rid="T1">Table 1</xref>. The ANOVA of mean amplitudes of 440&#x02013;600 ms showed a significant interaction between Load and Distractor type [<italic>F</italic><sub>(1, 31)</sub> = 4.69, <italic>p</italic> &#x0003C; 0.05, <inline-formula><mml:math id="M26"><mml:mrow><mml:msubsup><mml:mi>&#x003B7;</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:math></inline-formula> = 0.13; see <xref ref-type="fig" rid="F5">Figure 5</xref>]. Follow-up analyses showed that the mean amplitudes of (A + V) ERPs were more negative than those in the AV condition when the working memory load was high (<italic>M</italic> = &#x02212;0.88 vs. &#x02212;0.16 &#x003BC;V, SE = 0.38 vs. 0.25; <italic>t</italic><sub>(31)</sub> = 2.67, <italic>p</italic> &#x0003C; 0.05). However, under low load conditions, no significant differences in mean amplitudes were observed between these two conditions. Similarly, as in Experiment 2, the present results showed that audiovisual distractors were integrated under high load conditions but not under low load conditions. Moreover, the working memory load modulated the integration of audiovisual distractors at the late stage when the audiovisual distractors were the meaningful complex stimuli. Neither the main effect of Load [<italic>F</italic><sub>(1, 31)</sub> = 2.13, <italic>p</italic> > 0.05] nor the main effect of Distractor type [<italic>F</italic><sub>(1, 31)</sub> = 2.99, <italic>p</italic> > 0.05] was significant.</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p><bold>(A)</bold> Grand-average distractor-synchronized ERP waveforms at each experimental condition. Time window of 440&#x02013;600 ms is highlighted in the box (electrodes AF7, F5, F7). Time zero on the x-axis corresponds to the distractor stimulus onset. <bold>(B)</bold> Mean amplitudes and standard errors for different conditions during time window of 440&#x02013;600 ms (electrodes AF7, F5, F7). * indicates p &#x0003C; 0.05.</p></caption>
<graphic xlink:href="fnint-17-1120668-g0005.tif"/>
</fig>
<p>In addition, the ANOVAs of mean amplitudes of the early time window (250&#x02013;330 ms) showed no significant interaction between Load and Distractor type [<italic>F</italic><sub>(1, 31)</sub> = 0.26, <italic>p</italic> > 0.05]. Neither the main effect of Load [<italic>F</italic><sub>(1, 31)</sub> = 0.46, <italic>p</italic> > 0.05] nor the main effect of Distractor type [<italic>F</italic><sub>(1, 31)</sub> = 0.54, <italic>p</italic> > 0.05] was significant.</p>
</sec>
</sec>
<sec id="s4-3">
<title>4.3 Discussion</title>
<p>By adopting the ERP technique, the present results echo the main findings in Experiment 2. That is, the working memory load modulated the integration of audiovisual distractors. Specifically, audiovisual distractors could be integrated only under high load conditions. This was demonstrated by the significant interaction between Load and Distractor type, indicating there existed a late integration of AV distractors at the time window of 440&#x02013;600 ms over frontal electrodes.</p>
<p>Moreover, the complex distractor stimuli used in the present study were more ecological than the simple stimuli. As expected, a relative late integration effect of audiovisual distractors for complex stimuli was found, which was reflected by the late time window of significant integration effect in ERPs. This late integration is consistent with the findings in previous studies using complex audiovisual stimuli (Raij et al., <xref ref-type="bibr" rid="B40">2010</xref>; Xi et al., <xref ref-type="bibr" rid="B57">2020</xref>). For instance, Xi et al. (<xref ref-type="bibr" rid="B57">2020</xref>) adopted pictures and sounds of animals and inanimate objects as targets and asked participants to perform a discrimination task, in which participants had to attend to one side (left or right) and only respond to the target stimuli at that side. They found three late effects of semantic integration, that is, the time window of 220&#x02013;240 ms and 560&#x02013;600 ms for attended stimuli and the time window of 340&#x02013;360 ms for unattended stimuli. Together with the previous studies, our present results support that compared to the simple stimuli, it takes more time to integrate complex stimuli and thus a later time window of integration in ERPs was observed.</p>
</sec>
</sec>
<sec id="s5">
<title>5 General discussion</title>
<p>The present study aimed to investigate the effect of working memory load on the processing of audiovisual distractors. We measured the behavioral and electrophysiological responses to central letters with auditory, visual, and audiovisual distractors under different working memory loads. Through three experiments, we demonstrated that multisensory (audiovisual) distractors could effectively interfere with the n-back task, especially under high working memory load conditions. Moreover, the time course of the integration of audiovisual distractors depends on the complexity of distractor stimuli. That is, the integration of the audiovisual distractors is reflected as an early integration (240&#x02013;340 ms) of simple distractors and a late integration (440&#x02013;600 ms) of complex distractors.</p>
<p>First of all, behavioral performance in Experiment 1 showed slower responses to letters with peripheral audiovisual distractors than to letters without distractors, which was significant only under high working memory load condition. This result demonstrated that high working memory load strengthened the interference effects of audiovisual distractors. That is, audiovisual distractors were effective to interfere with the performance (e.g., slowing down the responses) in the main working memory task under high load condition. Similarly, our ERP results further supported this finding. In both Experiment 2 and Experiment 3, mean amplitudes were more pronounced for stimuli with (A + V) distractors than stimuli with AV distractors under high load conditions. By contrast, under low load conditions, no significant differences in mean amplitude were observed between these two conditions. These results indicate that working memory load modulates the integration of audiovisual distractors, which is consistent with previous results (de Fockert et al., <xref ref-type="bibr" rid="B11">2001</xref>). That is, in the low load condition, participants have enough attention resources to select and process task-relevant stimuli and inhibit the processing of distractors. However, in the high load condition, attention capacity is overloaded, leaving fewer resources to inhibit the distractors. Thus, the distractors disengage attention from the working memory task more easily under high load condition, i.e., showing the interference effect. These results suggest that compared to unisensory distractors, audiovisual distractors can be more effective in disengaging attention from the main working memory task, and the processing of audiovisual distractors is modulated by the availability of attention resources.</p>
<p>Moreover, ERPs results revealed both an early and a late integration of audiovisual distractors under high working memory load condition. Specifically, the integration of simple audiovisual distractors happened at the time window of 240&#x02013;340 ms after the onset of the distractors in Experiment 2, while a late integration of complex audiovisual distractors was found at the time window of 440&#x02013;600 ms in Experiment 3. Similar to Experiment 1, the integration of audiovisual distractors was only found under high load condition, suggesting that integration of audiovisual distractors needs top-down attention control. This finding is consistent with the integration framework of one early review (Koelewijn et al., <xref ref-type="bibr" rid="B24">2010</xref>). That is, unimodal inputs are processed independently in each modality and are then integrated at a late stage. Moreover, the different time courses of the multisensory integration of simple and complex distractors found in the current study might reflect the distinct awareness and processing of distractor stimuli. Compared with simple audiovisual distractors, the integration of complex audiovisual distractors may require semantic processing and is time-consuming, resulting in a late time window of integration. For instance, Xie et al. (<xref ref-type="bibr" rid="B58">2017</xref>) adopted a delayed matching-to-sample task, in which participants were required to judge whether the probe stimulus (visual) was the same as the target stimulus (visual, auditory, or audiovisual). The stimuli they used consisted of line drawings of real-life objects and the sound they made, such as animals, tools, vehicles, etc. They found a relatively late semantic target integration at the time window of 236&#x02013;530 ms, which was due to the requirement of top-down processing for the integration of semantic information.</p>
<p>Our study can broaden the understanding of the role of attention in multisensory integration. Previous studies in this field mainly focused on the crossmodal integration of targets (Santangelo and Spence, <xref ref-type="bibr" rid="B41">2007</xref>; Zimmer and Macaluso, <xref ref-type="bibr" rid="B59">2007</xref>). However, multisensory targets are supposed to be easily attended to and responded to. Thus, it is worth investigating the integration of multisensory distractors besides the targets, which can help us better understand whether attention is needed for multisensory integration. Here, we found that attention could modulate the audiovisual integration at both early and late stages, which is consistent with previous studies (Michail and Keil, <xref ref-type="bibr" rid="B35">2018</xref>; Lunn et al., <xref ref-type="bibr" rid="B30">2019</xref>) and frameworks trying to resolve the inconsistent results on the relationship between attention and multisensory integration (Koelewijn et al., <xref ref-type="bibr" rid="B24">2010</xref>; Navarra et al., <xref ref-type="bibr" rid="B37">2010</xref>; Talsma et al., <xref ref-type="bibr" rid="B49">2010</xref>). Researchers have considered key factors that modulate multisensory integration, such as stimuli complexity, stimuli competition, perceptual load, etc. Specifically, when the stimuli are complex, or the cognitive load is high, the current goal determines which stimuli are integrated first (top-down attentional control). When the stimuli are simple or the cognitive load is low, the stimuli could be integrated automatically (bottom-up processing without attention). Therefore, our results further demonstrate the importance of working memory load for the integration of audiovisual distractors.</p>
<p>It should be noted that although we found a significant interaction between Load and Distractor type in Experiment 1, i.e., significant interference effect of multisensory distractors under high WM load condition, such a behavioral result was not shown in Experiments 2 and 3. This inconsistency might be due to the slight difference across the experimental setting. On one hand, 7.5 degrees of visual angle was used for distractors in Experiment 1, while in Experiments 2 and 3, distractors were present at 4 degrees of visual angle. Previous studies have found that distractors could cause more interference at a peripheral location than at a central location (Chen, <xref ref-type="bibr" rid="B9">2008</xref>; Corral and Escera, <xref ref-type="bibr" rid="B10">2008</xref>), resulting in a decreased behavioral effect in both Experiments 2 and 3. On the other hand, a headphone was used in Experiment 1, while loudspeakers were adopted in Experiments 2 and 3. In Experiment 1, the audio was presented <italic>via</italic> headphones to the left or right ear of the participants. In Experiments 2 and 3, audio was presented <italic>via</italic> loudspeakers placed at the same location as the video behind the screen. Therefore, spatial (left or right side) information of audio in Experiments 2 and 3 was not as accurate as in Experiment 1, which might also reduce the interference effect of distractors. One previous study has reported a significant interference effect by peripheral sound distractors when using headphones but not using loudspeakers (Corral and Escera, <xref ref-type="bibr" rid="B10">2008</xref>). Nevertheless, we found significant multisensory integration effects in the ERP results of Experiments 2 and 3. Maybe the integration happened, but it was not strong enough to be observed at the behavioral level. For example, a previous study also used audiovisual distractors and found a significant early integration (around 50 ms) in ERP results but fail to observe the behavioral cost (Van der Burg et al., <xref ref-type="bibr" rid="B54">2011</xref>). Using unisensory and multisensory cues, Santangelo et al. (<xref ref-type="bibr" rid="B44">2008b</xref>) also revealed no increase orienting effect following bimodal as compared to unimodal cues, while the ERPs elicited by bimodal cues were more pronounced than the sum of the ERPs elicited by unisensory cues. These results suggest multisensory integration can happen even without observing behavioral benefits. In addition, the integration effect of distractors may not be as strong as that of targets shown in previous studies. Nevertheless, as the first evidence investigating the integration of distractors, we provide its cognitive and neural mechanisms by using the ERP method.</p>
<p>In conclusion, compared to unisensory auditory or visual distractors, multisensory audiovisual distractors can disengage participants&#x02019; attention more effectively, thus observing significant interference effects for audiovisual distractors. Moreover, working memory load modulates the processing of audiovisual distractors. Only under high load condition do the audiovisual distractors disengage attention from the working memory task and interfere with the task performance effectively. Our results support that attention is necessary for the occurrence of multisensory integration. Moreover, the integration of simple audiovisual distractors occurs at an early stage (240&#x02013;340 ms), while a late integration stage (440&#x02013;600 ms) for complex audiovisual distractors.</p>
</sec>
<sec id="s6" sec-type="data-availability">
<title>Data availability statement</title>
<p>The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found below: <ext-link ext-link-type="uri" xlink:href="https://osf.io/wh273">https://osf.io/wh273</ext-link>.</p>
</sec>
<sec id="s7" sec-type="ethics-statement">
<title>Ethics statement</title>
<p>The studies involving human participants were reviewed and approved and this study was approved by the Ethics Committee of Department of Psychology, Sun Yat-sen University (2020-0325-0127). The patients/participants provided their written informed consent to participate in this study.</p>
</sec>
<sec id="s8" sec-type="author-contributions">
<title>Author contributions</title>
<p>YY and XH: conceptualization, data curation, formal analysis, investigation, methodology, resources, software, validation, visualization, writing&#x02014;original draft, writing&#x02014;review and editing. ZY: conceptualization, funding acquisition, methodology, project administration, resources, supervision, writing&#x02014;original draft, writing&#x02014;review and editing. All authors contributed to the article and approved the submitted version.</p>
</sec>
</body>
<back>
<sec id="s9" sec-type="funding-information">
<title>Funding</title>
<p>The work was funded by the Fundamental Research Funds for the Central Universities, Sun Yat-sen University to ZY (22wklj04).</p>
</sec>
<sec id="s10" sec-type="COI-statement">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s11">
<title>Publisher&#x02019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Almadori</surname> <given-names>E.</given-names></name> <name><surname>Mastroberardino</surname> <given-names>S.</given-names></name> <name><surname>Botta</surname> <given-names>F.</given-names></name> <name><surname>Brunetti</surname> <given-names>R.</given-names></name> <name><surname>Lupi&#x000E1;&#x000F1;ez</surname> <given-names>J.</given-names></name> <name><surname>Spence</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Crossmodal semantic congruence interacts with object contextual consistency in complex visual scenes to enhance short-term memory performance</article-title>. <source>Brain Sci.</source> <volume>11</volume>:<fpage>1206</fpage>. <pub-id pub-id-type="doi">10.3390/brainsci11091206</pub-id><pub-id pub-id-type="pmid">34573227</pub-id></citation></ref>
<ref id="B2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alsius</surname> <given-names>A.</given-names></name> <name><surname>M&#x000F6;tt&#x000F6;nen</surname> <given-names>R.</given-names></name> <name><surname>Sams</surname> <given-names>M. E.</given-names></name> <name><surname>Soto-Faraco</surname> <given-names>S.</given-names></name> <name><surname>Tiippana</surname> <given-names>K.</given-names></name></person-group> (<year>2014</year>). <article-title>Effect of attentional load on audiovisual speech perception: evidence from ERPs</article-title>. <source>Front. Psychol.</source> <volume>5</volume>:<fpage>727</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2014.00727</pub-id><pub-id pub-id-type="pmid">25076922</pub-id></citation></ref>
<ref id="B3"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alsius</surname> <given-names>A.</given-names></name> <name><surname>Navarra</surname> <given-names>J.</given-names></name> <name><surname>Campbell</surname> <given-names>R.</given-names></name> <name><surname>Soto-Faraco</surname> <given-names>S.</given-names></name></person-group> (<year>2005</year>). <article-title>Audiovisual integration of speech falters under high attention demands</article-title>. <source>Curr. Biol.</source> <volume>15</volume>, <fpage>839</fpage>&#x02013;<lpage>843</lpage>. <pub-id pub-id-type="doi">10.1016/j.cub.2005.03.046</pub-id><pub-id pub-id-type="pmid">15886102</pub-id></citation></ref>
<ref id="B4"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bell</surname> <given-names>A. J.</given-names></name> <name><surname>Sejnowski</surname> <given-names>T. J.</given-names></name></person-group> (<year>1995</year>). <article-title>An information-maximization approach to blind separation and blind deconvolution</article-title>. <source>Neural Comput.</source> <volume>7</volume>, <fpage>1129</fpage>&#x02013;<lpage>1159</lpage>. <pub-id pub-id-type="doi">10.1162/neco.1995.7.6.1129</pub-id><pub-id pub-id-type="pmid">7584893</pub-id></citation></ref>
<ref id="B5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Botta</surname> <given-names>F.</given-names></name> <name><surname>Santangelo</surname> <given-names>V.</given-names></name> <name><surname>Raffone</surname> <given-names>A.</given-names></name> <name><surname>Lupi&#x000E1;&#x000F1;ez</surname> <given-names>J.</given-names></name> <name><surname>Belardinelli</surname> <given-names>M. O.</given-names></name></person-group> (<year>2010</year>). <article-title>Exogenous and endogenous spatial attention effects on visuospatial working memory</article-title>. <source>Q. J. Exp. Psychol. (Hove)</source> <volume>63</volume>, <fpage>1590</fpage>&#x02013;<lpage>1602</lpage>. <pub-id pub-id-type="doi">10.1080/17470210903443836</pub-id><pub-id pub-id-type="pmid">20112160</pub-id></citation></ref>
<ref id="B6"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Botta</surname> <given-names>F.</given-names></name> <name><surname>Santangelo</surname> <given-names>V.</given-names></name> <name><surname>Raffone</surname> <given-names>A.</given-names></name> <name><surname>Sanabria</surname> <given-names>D.</given-names></name> <name><surname>Lupi&#x000E1;&#x000F1;ez</surname> <given-names>J.</given-names></name> <name><surname>Belardinelli</surname> <given-names>M. O.</given-names></name></person-group> (<year>2011</year>). <article-title>Multisensory integration affects visuo-spatial working memory</article-title>. <source>J. Exp. Psychol. Hum. Percept. Perform.</source> <volume>37</volume>, <fpage>1099</fpage>&#x02013;<lpage>1109</lpage>. <pub-id pub-id-type="doi">10.1037/a0023513</pub-id><pub-id pub-id-type="pmid">21553989</pub-id></citation></ref>
<ref id="B7"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brunetti</surname> <given-names>R.</given-names></name> <name><surname>Indraccolo</surname> <given-names>A.</given-names></name> <name><surname>Mastroberardino</surname> <given-names>S.</given-names></name> <name><surname>Spence</surname> <given-names>C.</given-names></name> <name><surname>Santangelo</surname> <given-names>V.</given-names></name></person-group> (<year>2017</year>). <article-title>The impact of cross-modal correspondences on working memory performance</article-title>. <source>J. Exp. Psychol. Hum. Percept. Perform.</source> <volume>43</volume>, <fpage>819</fpage>&#x02013;<lpage>831</lpage>. <pub-id pub-id-type="doi">10.1037/xhp0000348</pub-id><pub-id pub-id-type="pmid">28345948</pub-id></citation></ref>
<ref id="B8"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Caclin</surname> <given-names>A.</given-names></name> <name><surname>Soto-Faraco</surname> <given-names>S.</given-names></name> <name><surname>Kingstone</surname> <given-names>A.</given-names></name> <name><surname>Spence</surname> <given-names>C.</given-names></name></person-group> (<year>2002</year>). <article-title>Tactile &#x0201C;capture&#x0201D; of audition</article-title>. <source>Percept. Psychophys.</source> <volume>64</volume>, <fpage>616</fpage>&#x02013;<lpage>630</lpage>. <pub-id pub-id-type="doi">10.3758/bf03194730</pub-id><pub-id pub-id-type="pmid">12132762</pub-id></citation></ref>
<ref id="B9"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>Z.</given-names></name></person-group> (<year>2008</year>). <article-title>Distractor eccentricity and its effect on selective attention</article-title>. <source>Exp. Psychol.</source> <volume>55</volume>, <fpage>82</fpage>&#x02013;<lpage>92</lpage>. <pub-id pub-id-type="doi">10.1027/1618-3169.55.2.82</pub-id><pub-id pub-id-type="pmid">18444518</pub-id></citation></ref>
<ref id="B10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Corral</surname> <given-names>M. J.</given-names></name> <name><surname>Escera</surname> <given-names>C.</given-names></name></person-group> (<year>2008</year>). <article-title>Effects of sound location on visual task performance and electrophysiological measures of distraction</article-title>. <source>Neuroreport</source> <volume>19</volume>, <fpage>1535</fpage>&#x02013;<lpage>1539</lpage>. <pub-id pub-id-type="doi">10.1097/WNR.0b013e3283110416</pub-id><pub-id pub-id-type="pmid">18797312</pub-id></citation></ref>
<ref id="B11"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>de Fockert</surname> <given-names>J. W.</given-names></name> <name><surname>Rees</surname> <given-names>G.</given-names></name> <name><surname>Frith</surname> <given-names>C. D.</given-names></name> <name><surname>Lavie</surname> <given-names>N.</given-names></name></person-group> (<year>2001</year>). <article-title>The role of working memory in visual selective attention</article-title>. <source>Science</source> <volume>291</volume>, <fpage>1803</fpage>&#x02013;<lpage>1806</lpage>. <pub-id pub-id-type="doi">10.1126/science.1056496</pub-id><pub-id pub-id-type="pmid">11230699</pub-id></citation></ref>
<ref id="B13"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Diederich</surname> <given-names>A.</given-names></name> <name><surname>Colonius</surname> <given-names>H.</given-names></name></person-group> (<year>2004</year>). <article-title>Bimodal and trimodal multisensory enhancement: effects of stimulus onset and intensity on reaction time</article-title>. <source>Percept. Psychophys.</source> <volume>66</volume>, <fpage>1388</fpage>&#x02013;<lpage>1404</lpage>. <pub-id pub-id-type="doi">10.3758/bf03195006</pub-id><pub-id pub-id-type="pmid">15813202</pub-id></citation></ref>
<ref id="B14"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Downing</surname> <given-names>P. E.</given-names></name></person-group> (<year>2000</year>). <article-title>Interactions between visual working memory and selective attention</article-title>. <source>Psychol. Sci.</source> <volume>11</volume>, <fpage>467</fpage>&#x02013;<lpage>473</lpage>. <pub-id pub-id-type="doi">10.1111/1467-9280.00290</pub-id><pub-id pub-id-type="pmid">11202491</pub-id></citation></ref>
<ref id="B16"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Faul</surname> <given-names>F.</given-names></name> <name><surname>Erdfelder</surname> <given-names>E.</given-names></name> <name><surname>Buchner</surname> <given-names>A.</given-names></name> <name><surname>Lang</surname> <given-names>A. G.</given-names></name></person-group> (<year>2009</year>). <article-title>Statistical power analyses using G*Power 3.1: tests for correlation and regression analyses</article-title>. <source>Behav. Res. Methods</source> <volume>41</volume>, <fpage>1149</fpage>&#x02013;<lpage>1160</lpage>. <pub-id pub-id-type="doi">10.3758/BRM.41.4.1149</pub-id><pub-id pub-id-type="pmid">19897823</pub-id></citation></ref>
<ref id="B17"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Frassinetti</surname> <given-names>F.</given-names></name> <name><surname>Bolognini</surname> <given-names>N.</given-names></name> <name><surname>L&#x000E0;davas</surname> <given-names>E.</given-names></name></person-group> (<year>2002</year>). <article-title>Enhancement of visual perception by crossmodal visuo-auditory interaction</article-title>. <source>Exp. Brain Res.</source> <volume>147</volume>, <fpage>332</fpage>&#x02013;<lpage>343</lpage>. <pub-id pub-id-type="doi">10.1007/s00221-002-1262-y</pub-id><pub-id pub-id-type="pmid">12428141</pub-id></citation></ref>
<ref id="B18"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Giard</surname> <given-names>M. H.</given-names></name> <name><surname>Peronnet</surname> <given-names>F.</given-names></name></person-group> (<year>1999</year>). <article-title>Auditory-visual integration during multimodal object recognition in humans: a behavioral and electrophysiological study</article-title>. <source>J. Cogn. Neurosci.</source> <volume>11</volume>, <fpage>473</fpage>&#x02013;<lpage>490</lpage>. <pub-id pub-id-type="doi">10.1162/089892999563544</pub-id><pub-id pub-id-type="pmid">10511637</pub-id></citation></ref>
<ref id="B19"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gibney</surname> <given-names>K. D.</given-names></name> <name><surname>Aligbe</surname> <given-names>E.</given-names></name> <name><surname>Eggleston</surname> <given-names>B. A.</given-names></name> <name><surname>Nunes</surname> <given-names>S. R.</given-names></name> <name><surname>Kerkhoff</surname> <given-names>W. G.</given-names></name> <name><surname>Dean</surname> <given-names>C. L.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>Visual distractors disrupt audiovisual integration regardless of stimulus complexity</article-title>. <source>Front. Integr. Neurosci.</source> <volume>11</volume>:<fpage>1</fpage>. <pub-id pub-id-type="doi">10.3389/fnint.2017.00001</pub-id><pub-id pub-id-type="pmid">28163675</pub-id></citation></ref>
<ref id="B20"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hyun</surname> <given-names>J. S.</given-names></name> <name><surname>Woodman</surname> <given-names>G. F.</given-names></name> <name><surname>Luck</surname> <given-names>S. J.</given-names></name></person-group> (<year>2009</year>). <article-title>The role of attention in the binding of surface features to locations</article-title>. <source>Vis. Cogn.</source> <volume>17</volume>:<fpage>10.1080/13506280802113894</fpage>. <pub-id pub-id-type="doi">10.1080/13506280802113894</pub-id><pub-id pub-id-type="pmid">24235876</pub-id></citation></ref>
<ref id="B21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jensen</surname> <given-names>A.</given-names></name> <name><surname>Merz</surname> <given-names>S.</given-names></name> <name><surname>Spence</surname> <given-names>C.</given-names></name> <name><surname>Frings</surname> <given-names>C.</given-names></name></person-group> (<year>2019</year>). <article-title>Overt spatial attention modulates multisensory selection</article-title>. <source>J. Exp. Psychol. Hum. Percept. Perform.</source> <volume>45</volume>, <fpage>174</fpage>&#x02013;<lpage>188</lpage>. <pub-id pub-id-type="doi">10.1037/xhp0000595</pub-id><pub-id pub-id-type="pmid">30589358</pub-id></citation></ref>
<ref id="B22"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jensen</surname> <given-names>A.</given-names></name> <name><surname>Merz</surname> <given-names>S.</given-names></name> <name><surname>Spence</surname> <given-names>C.</given-names></name> <name><surname>Frings</surname> <given-names>C.</given-names></name></person-group> (<year>2020</year>). <article-title>Interference of irrelevant information in multisensory selection depends on attentional set</article-title>. <source>Attent. Percept. Psychophys.</source> <volume>82</volume>, <fpage>1176</fpage>&#x02013;<lpage>1195</lpage>. <pub-id pub-id-type="doi">10.3758/s13414-019-01848-8</pub-id><pub-id pub-id-type="pmid">31444699</pub-id></citation></ref>
<ref id="B23"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Keil</surname> <given-names>J.</given-names></name> <name><surname>Senkowski</surname> <given-names>D.</given-names></name></person-group> (<year>2018</year>). <article-title>Neural oscillations orchestrate multisensory processing</article-title>. <source>Neuroscientist</source> <volume>24</volume>, <fpage>609</fpage>&#x02013;<lpage>626</lpage>. <pub-id pub-id-type="doi">10.1177/1073858418755352</pub-id><pub-id pub-id-type="pmid">29424265</pub-id></citation></ref>
<ref id="B24"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Koelewijn</surname> <given-names>T.</given-names></name> <name><surname>Bronkhorst</surname> <given-names>A.</given-names></name> <name><surname>Theeuwes</surname> <given-names>J.</given-names></name></person-group> (<year>2010</year>). <article-title>Attention and the multiple stages of multisensory integration: a review of audiovisual studies</article-title>. <source>Acta Psychol. (Amst)</source> <volume>134</volume>, <fpage>372</fpage>&#x02013;<lpage>384</lpage>. <pub-id pub-id-type="doi">10.1016/j.actpsy.2010.03.010</pub-id><pub-id pub-id-type="pmid">20427031</pub-id></citation></ref>
<ref id="B25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kvasova</surname> <given-names>D.</given-names></name> <name><surname>Garcia-Vernet</surname> <given-names>L.</given-names></name> <name><surname>Soto-Faraco</surname> <given-names>S.</given-names></name></person-group> (<year>2019</year>). <article-title>Characteristic sounds facilitate object search in real-life scenes</article-title>. <source>Front. Psychol.</source> <volume>10</volume>:<fpage>2511</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2019.02511</pub-id><pub-id pub-id-type="pmid">31749751</pub-id></citation></ref>
<ref id="B30"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lunn</surname> <given-names>J.</given-names></name> <name><surname>Sjoblom</surname> <given-names>A.</given-names></name> <name><surname>Ward</surname> <given-names>J.</given-names></name> <name><surname>Soto-Faraco</surname> <given-names>S.</given-names></name> <name><surname>Forster</surname> <given-names>S.</given-names></name></person-group> (<year>2019</year>). <article-title>Multisensory enhancement of attention depends on whether you are already paying attention</article-title>. <source>Cognition</source> <volume>187</volume>, <fpage>38</fpage>&#x02013;<lpage>49</lpage>. <pub-id pub-id-type="doi">10.1016/j.cognition.2019.02.008</pub-id><pub-id pub-id-type="pmid">30825813</pub-id></citation></ref>
<ref id="B32"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mastroberardino</surname> <given-names>S.</given-names></name> <name><surname>Santangelo</surname> <given-names>V.</given-names></name> <name><surname>Botta</surname> <given-names>F.</given-names></name> <name><surname>Marucci</surname> <given-names>F. S.</given-names></name> <name><surname>Olivetti Belardinelli</surname> <given-names>M.</given-names></name></person-group> (<year>2008</year>). <article-title>How the bimodal format of presentation affects working memory: an overview</article-title>. <source>Cogn. Process.</source> <volume>9</volume>, <fpage>69</fpage>&#x02013;<lpage>76</lpage>. <pub-id pub-id-type="doi">10.1007/s10339-007-0195-6</pub-id><pub-id pub-id-type="pmid">17932697</pub-id></citation></ref>
<ref id="B31"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mastroberardino</surname> <given-names>S.</given-names></name> <name><surname>Santangelo</surname> <given-names>V.</given-names></name> <name><surname>Macaluso</surname> <given-names>E.</given-names></name></person-group> (<year>2015</year>). <article-title>Crossmodal semantic congruence can affect visuo-spatial processing and activity of the fronto-parietal attention networks</article-title>. <source>Front. Integr. Neurosci.</source> <volume>9</volume>:<fpage>45</fpage>. <pub-id pub-id-type="doi">10.3389/fnint.2015.00045</pub-id><pub-id pub-id-type="pmid">26217199</pub-id></citation></ref>
<ref id="B33"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Merz</surname> <given-names>S.</given-names></name> <name><surname>Frings</surname> <given-names>C.</given-names></name> <name><surname>Spence</surname> <given-names>C.</given-names></name></person-group> (<year>2021</year>). <article-title>When irrelevant information helps: extending the Eriksen-flanker task into a multisensory world</article-title>. <source>Attent. Percept. Psychophys.</source> <volume>83</volume>, <fpage>776</fpage>&#x02013;<lpage>789</lpage>. <pub-id pub-id-type="doi">10.3758/s13414-020-02066-3</pub-id><pub-id pub-id-type="pmid">32514664</pub-id></citation></ref>
<ref id="B34"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Merz</surname> <given-names>S.</given-names></name> <name><surname>Jensen</surname> <given-names>A.</given-names></name> <name><surname>Spence</surname> <given-names>C.</given-names></name> <name><surname>Frings</surname> <given-names>C.</given-names></name></person-group> (<year>2019</year>). <article-title>Multisensory distractor processing is modulated by spatial attention</article-title>. <source>J. Exp. Psychol. Hum. Percept. Perform.</source> <volume>45</volume>, <fpage>1375</fpage>&#x02013;<lpage>1388</lpage>. <pub-id pub-id-type="doi">10.1037/xhp0000678</pub-id><pub-id pub-id-type="pmid">31343245</pub-id></citation></ref>
<ref id="B35"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Michail</surname> <given-names>G.</given-names></name> <name><surname>Keil</surname> <given-names>J.</given-names></name></person-group> (<year>2018</year>). <article-title>High cognitive load enhances the susceptibility to non-speech audiovisual illusions</article-title>. <source>Sci. Rep.</source> <volume>8</volume>:<fpage>11530</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-018-30007-6</pub-id><pub-id pub-id-type="pmid">30069059</pub-id></citation></ref>
<ref id="B37"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Navarra</surname> <given-names>J.</given-names></name> <name><surname>Alsius</surname> <given-names>A.</given-names></name> <name><surname>Soto-Faraco</surname> <given-names>S.</given-names></name> <name><surname>Spence</surname> <given-names>C.</given-names></name></person-group> (<year>2010</year>). <article-title>Assessing the role of attention in the audiovisual integration of speech</article-title>. <source>Inform. Fusion</source> <volume>11</volume>, <fpage>4</fpage>&#x02013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1016/j.inffus.2009.04.001</pub-id></citation></ref>
<ref id="B38"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Oberauer</surname> <given-names>K.</given-names></name></person-group> (<year>2019</year>). <article-title>Working memory and attention - a conceptual analysis and review</article-title>. <source>J. Cogn.</source> <volume>2</volume>:<fpage>36</fpage>. <pub-id pub-id-type="doi">10.5334/joc.58</pub-id><pub-id pub-id-type="pmid">31517246</pub-id></citation></ref>
<ref id="B40"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Raij</surname> <given-names>T.</given-names></name> <name><surname>Ahveninen</surname> <given-names>J.</given-names></name> <name><surname>Lin</surname> <given-names>F. H.</given-names></name> <name><surname>Witzel</surname> <given-names>T.</given-names></name> <name><surname>J&#x000E4;&#x000E4;skel&#x000E4;inen</surname> <given-names>I. P.</given-names></name> <name><surname>Letham</surname> <given-names>B.</given-names></name> <etal/></person-group>. (<year>2010</year>). <article-title>Onset timing of cross-sensory activations and multisensory interactions in auditory and visual sensory cortices</article-title>. <source>Eur. J. Neurosci.</source> <volume>31</volume>, <fpage>1772</fpage>&#x02013;<lpage>1782</lpage>. <pub-id pub-id-type="doi">10.1111/j.1460-9568.2010.07213.x</pub-id><pub-id pub-id-type="pmid">20584181</pub-id></citation></ref>
<ref id="B42"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Santangelo</surname> <given-names>V.</given-names></name> <name><surname>Ho</surname> <given-names>C.</given-names></name> <name><surname>Spence</surname> <given-names>C.</given-names></name></person-group> (<year>2008a</year>). <article-title>Capturing spatial attention with multisensory cues</article-title>. <source>Psychon. Bull. Rev.</source> <volume>15</volume>, <fpage>398</fpage>&#x02013;<lpage>403</lpage>. <pub-id pub-id-type="doi">10.3758/pbr.15.2.398</pub-id><pub-id pub-id-type="pmid">18488658</pub-id></citation></ref>
<ref id="B41"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Santangelo</surname> <given-names>V.</given-names></name> <name><surname>Spence</surname> <given-names>C.</given-names></name></person-group> (<year>2007</year>). <article-title>Multisensory cues capture spatial attention regardless of perceptual load</article-title>. <source>J. Exp. Psychol. Hum. Percept. Perform.</source> <volume>33</volume>, <fpage>1311</fpage>&#x02013;<lpage>1321</lpage>. <pub-id pub-id-type="doi">10.1037/0096-1523.33.6.1311</pub-id><pub-id pub-id-type="pmid">18085945</pub-id></citation></ref>
<ref id="B43"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Santangelo</surname> <given-names>V.</given-names></name> <name><surname>Van der Lubbe</surname> <given-names>R. H.</given-names></name> <name><surname>Belardinelli</surname> <given-names>M. O.</given-names></name> <name><surname>Postma</surname> <given-names>A.</given-names></name></person-group> (<year>2006</year>). <article-title>Spatial attention triggered by unimodal, crossmodal and bimodal exogenous cues: a comparison of reflexive orienting mechanisms</article-title>. <source>Exp. Brain Res.</source> <volume>173</volume>, <fpage>40</fpage>&#x02013;<lpage>48</lpage>. <pub-id pub-id-type="doi">10.1007/s00221-006-0361-6</pub-id><pub-id pub-id-type="pmid">16489435</pub-id></citation></ref>
<ref id="B44"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Santangelo</surname> <given-names>V.</given-names></name> <name><surname>Van der Lubbe</surname> <given-names>R. H.</given-names></name> <name><surname>Olivetti Belardinelli</surname> <given-names>M.</given-names></name> <name><surname>Postma</surname> <given-names>A.</given-names></name></person-group> (<year>2008b</year>). <article-title>Multisensory integration affects ERP components elicited by exogenous cues</article-title>. <source>Exp. Brain Res.</source> <volume>185</volume>, <fpage>269</fpage>&#x02013;<lpage>277</lpage>. <pub-id pub-id-type="doi">10.1007/s00221-007-1151-5</pub-id><pub-id pub-id-type="pmid">17909764</pub-id></citation></ref>
<ref id="B45"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stein</surname> <given-names>B. E.</given-names></name> <name><surname>Burr</surname> <given-names>D.</given-names></name> <name><surname>Constantinidis</surname> <given-names>C.</given-names></name> <name><surname>Laurienti</surname> <given-names>P. J.</given-names></name> <name><surname>Alex Meredith</surname> <given-names>M.</given-names></name> <name><surname>Perrault</surname> <given-names>T. J.</given-names></name> <etal/></person-group>. (<year>2010</year>). <article-title>Semantic confusion regarding the development of multisensory integration: a practical solution</article-title>. <source>Eur. J. Neurosci.</source> <volume>31</volume>, <fpage>1713</fpage>&#x02013;<lpage>1720</lpage>. <pub-id pub-id-type="doi">10.1111/j.1460-9568.2010.07206.x</pub-id><pub-id pub-id-type="pmid">20584174</pub-id></citation></ref>
<ref id="B46"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stevenson</surname> <given-names>R. A.</given-names></name> <name><surname>Ghose</surname> <given-names>D.</given-names></name> <name><surname>Fister</surname> <given-names>J. K.</given-names></name> <name><surname>Sarko</surname> <given-names>D. K.</given-names></name> <name><surname>Altieri</surname> <given-names>N. A.</given-names></name> <name><surname>Nidiffer</surname> <given-names>A. R.</given-names></name> <etal/></person-group>. (<year>2014</year>). <article-title>Identifying and quantifying multisensory integration: a tutorial review</article-title>. <source>Brain Topogr.</source> <volume>27</volume>, <fpage>707</fpage>&#x02013;<lpage>730</lpage>. <pub-id pub-id-type="doi">10.1007/s10548-014-0365-7</pub-id><pub-id pub-id-type="pmid">24722880</pub-id></citation></ref>
<ref id="B48"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Talsma</surname> <given-names>D.</given-names></name> <name><surname>Doty</surname> <given-names>T. J.</given-names></name> <name><surname>Woldorff</surname> <given-names>M. G.</given-names></name></person-group> (<year>2007</year>). <article-title>Selective attention and audiovisual integration: is attending to both modalities a prerequisite for early integration?</article-title> <source>Cereb. Cortex</source> <volume>17</volume>, <fpage>679</fpage>&#x02013;<lpage>690</lpage>. <pub-id pub-id-type="doi">10.1093/cercor/bhk016</pub-id><pub-id pub-id-type="pmid">16707740</pub-id></citation></ref>
<ref id="B49"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Talsma</surname> <given-names>D.</given-names></name> <name><surname>Senkowski</surname> <given-names>D.</given-names></name> <name><surname>Soto-Faraco</surname> <given-names>S.</given-names></name> <name><surname>Woldorff</surname> <given-names>M. G.</given-names></name></person-group> (<year>2010</year>). <article-title>The multifaceted interplay between attention and multisensory integration</article-title>. <source>Trends Cogn. Sci.</source> <volume>14</volume>, <fpage>400</fpage>&#x02013;<lpage>410</lpage>. <pub-id pub-id-type="doi">10.1016/j.tics.2010.06.008</pub-id><pub-id pub-id-type="pmid">20675182</pub-id></citation></ref>
<ref id="B47"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Talsma</surname> <given-names>D.</given-names></name> <name><surname>Woldorff</surname> <given-names>M. G.</given-names></name></person-group> (<year>2005</year>). <article-title>Selective attention and multisensory integration: multiple phases of effects on the evoked brain activity</article-title>. <source>J. Cogn. Neurosci.</source> <volume>17</volume>, <fpage>1098</fpage>&#x02013;<lpage>1114</lpage>. <pub-id pub-id-type="doi">10.1162/0898929054475172</pub-id><pub-id pub-id-type="pmid">16102239</pub-id></citation></ref>
<ref id="B50"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Teder-S&#x000E4;lej&#x000E4;rvi</surname> <given-names>W. A.</given-names></name> <name><surname>McDonald</surname> <given-names>J. J.</given-names></name> <name><surname>Di Russo</surname> <given-names>F.</given-names></name> <name><surname>Hillyard</surname> <given-names>S. A.</given-names></name></person-group> (<year>2002</year>). <article-title>An analysis of audio-visual crossmodal integration by means of event-related potential (ERP) recordings</article-title>. <source>Cogn. Brain Res.</source> <volume>14</volume>, <fpage>106</fpage>&#x02013;<lpage>114</lpage>. <pub-id pub-id-type="doi">10.1016/s0926-6410(02)00065-4</pub-id><pub-id pub-id-type="pmid">12063134</pub-id></citation></ref>
<ref id="B51"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>ten Oever</surname> <given-names>S.</given-names></name> <name><surname>Romei</surname> <given-names>V.</given-names></name> <name><surname>van Atteveldt</surname> <given-names>N.</given-names></name> <name><surname>Soto-Faraco</surname> <given-names>S.</given-names></name> <name><surname>Murray</surname> <given-names>M. M.</given-names></name> <name><surname>Matusz</surname> <given-names>P. J.</given-names></name></person-group> (<year>2016</year>). <article-title>The COGs (context, object and goals) in multisensory processing</article-title>. <source>Exp. Brain Res.</source> <volume>234</volume>, <fpage>1307</fpage>&#x02013;<lpage>1323</lpage>. <pub-id pub-id-type="doi">10.1007/s00221-016-4590-z</pub-id><pub-id pub-id-type="pmid">26931340</pub-id></citation></ref>
<ref id="B52"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van der Burg</surname> <given-names>E.</given-names></name> <name><surname>Olivers</surname> <given-names>C. N.</given-names></name> <name><surname>Bronkhorst</surname> <given-names>A. W.</given-names></name> <name><surname>Theeuwes</surname> <given-names>J.</given-names></name></person-group> (<year>2008</year>). <article-title>Pip and pop: nonspatial auditory signals improve spatial visual search</article-title>. <source>J. Exp. Psychol. Hum. Percept. Perform.</source> <volume>34</volume>, <fpage>1053</fpage>&#x02013;<lpage>1065</lpage>. <pub-id pub-id-type="doi">10.1037/0096-1523.34.5.1053</pub-id><pub-id pub-id-type="pmid">18823194</pub-id></citation></ref>
<ref id="B53"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van der Burg</surname> <given-names>E.</given-names></name> <name><surname>Olivers</surname> <given-names>C. N.</given-names></name> <name><surname>Bronkhorst</surname> <given-names>A. W.</given-names></name> <name><surname>Theeuwes</surname> <given-names>J.</given-names></name></person-group> (<year>2009</year>). <article-title>Poke and pop: tactile-visual synchrony increases visual saliency</article-title>. <source>Neurosci. Lett.</source> <volume>450</volume>, <fpage>60</fpage>&#x02013;<lpage>64</lpage>. <pub-id pub-id-type="doi">10.1016/j.neulet.2008.11.002</pub-id><pub-id pub-id-type="pmid">19013216</pub-id></citation></ref>
<ref id="B54"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van der Burg</surname> <given-names>E.</given-names></name> <name><surname>Talsma</surname> <given-names>D.</given-names></name> <name><surname>Olivers</surname> <given-names>C. N.</given-names></name> <name><surname>Hickey</surname> <given-names>C.</given-names></name> <name><surname>Theeuwes</surname> <given-names>J.</given-names></name></person-group> (<year>2011</year>). <article-title>Early multisensory interactions affect the competition among multiple visual objects</article-title>. <source>Neuroimage</source> <volume>55</volume>, <fpage>1208</fpage>&#x02013;<lpage>1218</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2010.12.068</pub-id><pub-id pub-id-type="pmid">21195781</pub-id></citation></ref>
<ref id="B55"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van der Stoep</surname> <given-names>N.</given-names></name> <name><surname>Nijboer</surname> <given-names>T. C.</given-names></name> <name><surname>Van der Stigchel</surname> <given-names>S.</given-names></name></person-group> (<year>2014</year>). <article-title>Exogenous orienting of crossmodal attention in 3-D space: support for a depth-aware crossmodal attentional system</article-title>. <source>Psychon. Bull. Rev.</source> <volume>21</volume>, <fpage>708</fpage>&#x02013;<lpage>714</lpage>. <pub-id pub-id-type="doi">10.3758/s13423-013-0532-y</pub-id><pub-id pub-id-type="pmid">24101573</pub-id></citation></ref>
<ref id="B56"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wahn</surname> <given-names>B.</given-names></name> <name><surname>Murali</surname> <given-names>S.</given-names></name> <name><surname>Sinnett</surname> <given-names>S.</given-names></name> <name><surname>K&#x000F6;nig</surname> <given-names>P.</given-names></name></person-group> (<year>2017</year>). <article-title>Auditory stimulus detection partially depends on visuospatial attentional resources</article-title>. <source>Iperception</source> <volume>8</volume>:<fpage>2041669516688026</fpage>. <pub-id pub-id-type="doi">10.1177/2041669516688026</pub-id><pub-id pub-id-type="pmid">28203353</pub-id></citation></ref>
<ref id="B57"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xi</surname> <given-names>Y.</given-names></name> <name><surname>Li</surname> <given-names>Q.</given-names></name> <name><surname>Gao</surname> <given-names>N.</given-names></name> <name><surname>Li</surname> <given-names>G.</given-names></name> <name><surname>Lin</surname> <given-names>W.</given-names></name> <name><surname>Wu</surname> <given-names>J.</given-names></name></person-group> (<year>2020</year>). <article-title>Co-stimulation-removed audiovisual semantic integration and modulation of attention: an event-related potential study</article-title>. <source>Int. J. Psychophysiol.</source> <volume>151</volume>, <fpage>7</fpage>&#x02013;<lpage>17</lpage>. <pub-id pub-id-type="doi">10.1016/j.ijpsycho.2020.02.009</pub-id><pub-id pub-id-type="pmid">32061614</pub-id></citation></ref>
<ref id="B58"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xie</surname> <given-names>Y.</given-names></name> <name><surname>Xu</surname> <given-names>Y.</given-names></name> <name><surname>Bian</surname> <given-names>C.</given-names></name> <name><surname>Li</surname> <given-names>M.</given-names></name></person-group> (<year>2017</year>). <article-title>Semantic congruent audiovisual integration during the encoding stage of working memory: an ERP and sLORETA study</article-title>. <source>Sci. Rep.</source> <volume>7</volume>:<fpage>5112</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-017-05471-1</pub-id><pub-id pub-id-type="pmid">28698594</pub-id></citation></ref>
<ref id="B59"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zimmer</surname> <given-names>U.</given-names></name> <name><surname>Macaluso</surname> <given-names>E.</given-names></name></person-group> (<year>2007</year>). <article-title>Processing of multisensory spatial congruency can be dissociated from working memory and visuo-spatial attention</article-title>. <source>Eur. J. Neurosci.</source> <volume>26</volume>, <fpage>1681</fpage>&#x02013;<lpage>1691</lpage>. <pub-id pub-id-type="doi">10.1111/j.1460-9568.2007.05784.x</pub-id><pub-id pub-id-type="pmid">17880400</pub-id></citation></ref>
</ref-list>
<fn-group>
<fn id="fn0001"><p><sup>1</sup><ext-link ext-link-type="uri" xlink:href="http://www.pstnet.com/">http://www.pstnet.com/</ext-link></p></fn>
<fn id="fn0002"><p><sup>2</sup><ext-link ext-link-type="uri" xlink:href="https://www.mathworks.com/">https://www.mathworks.com/</ext-link></p></fn>
</fn-group>
</back>
</article>
