<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Hum. Neurosci.</journal-id>
<journal-title>Frontiers in Human Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Hum. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-5161</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnhum.2021.725449</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Weighted Integration of Duration Information Across Visual and Auditory Modality Is Influenced by Modality-Specific Attention</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Yoshimatsu</surname> <given-names>Hiroshi</given-names></name>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Yotsumoto</surname> <given-names>Yuko</given-names></name>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/75651/overview"/>
</contrib>
</contrib-group>
<aff><institution>Department of Life Sciences, The University of Tokyo</institution>, <addr-line>Tokyo</addr-line>, <country>Japan</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Filippo Brighina, University of Palermo, Italy</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Yasuki Noguchi, Kobe University, Japan; Aysha Motala, Western University, Canada</p></fn>
<corresp id="c001">&#x002A;Correspondence: Yuko Yotsumoto, <email>cyuko@mail.ecc.u-tokyo.ac.jp</email></corresp>
<fn fn-type="other" id="fn004"><p>This article was submitted to Sensory Neuroscience, a section of the journal Frontiers in Human Neuroscience</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>07</day>
<month>10</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>15</volume>
<elocation-id>725449</elocation-id>
<history>
<date date-type="received">
<day>17</day>
<month>06</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>13</day>
<month>09</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2021 Yoshimatsu and Yotsumoto.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Yoshimatsu and Yotsumoto</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>We constantly integrate multiple types of information from different sensory modalities. Generally, such integration is influenced by the modality that we attend to. However, for duration perception, it has been shown that when duration information from visual and auditory modalities is integrated, the perceived duration of the visual stimulus leaned toward the duration of the auditory stimulus, irrespective of which modality was attended. In these studies, auditory dominance was assessed using visual and auditory stimuli with different durations whose timing of onset and offset would affect perception. In the present study, we aimed to investigate the effect of attention on duration integration using visual and auditory stimuli of the same duration. Since the duration of a visual flicker and auditory flutter tends to be perceived as longer than and shorter than its physical duration, respectively, we used the 10 Hz visual flicker and auditory flutter with the same onset and offset timings but different perceived durations. The participants were asked to attend either visual, auditory, or both modalities. Contrary to the attention-independent auditory dominance reported in previous studies, we found that the perceived duration of the simultaneous flicker and flutter presentation depended on which modality the participants attended. To further investigate the process of duration integration of the two modalities, we applied Bayesian hierarchical modeling, which enabled us to define a flexible model in which the multisensory duration is represented by the weighted average of each sensory modality. In addition, to examine whether auditory dominance results from the higher reliability of auditory stimuli, we applied another models to consider the stimulus reliability. These behavioral and modeling results suggest the following: (1) the perceived duration of visual and auditory stimuli is influenced by which modality the participants attended to when we control for the confounding effect of onset&#x2013;offset timing of stimuli, and (2) the increase of the weight by attention affects the duration integration, even when the effect of stimulus reliability is controlled. Our models can be extended to investigate the neural basis and effects of other sensory modalities in duration integration.</p>
</abstract>
<kwd-group>
<kwd>time perception</kwd>
<kwd>duration perception</kwd>
<kwd>multisensory integration</kwd>
<kwd>modality-specific attention</kwd>
<kwd>Bayesian hierarchical model</kwd>
</kwd-group>
<counts>
<fig-count count="10"/>
<table-count count="2"/>
<equation-count count="10"/>
<ref-count count="37"/>
<page-count count="14"/>
<word-count count="8944"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="S1">
<title>Introduction</title>
<p>Our perception results from receiving multimodal inputs from one&#x2019;s surroundings and integrating these inputs as one. However, integrated multimodal perception may not reflect the true reality. For example, the ventriloquism effect is the illusory percept, where the perceived location of an auditory stimulus is influenced by the location of the simultaneously presented visual stimulus (<xref ref-type="bibr" rid="B2">Battaglia et al., 2003</xref>; <xref ref-type="bibr" rid="B1">Alais and Burr, 2004</xref>). Similarly, this illusion is also found in the temporal domain, as the temporal ventriloquism effect, in which the perceived timing of a visual stimulus is drawn to the timing of the auditory stimulus (<xref ref-type="bibr" rid="B19">Klink et al., 2011</xref>; <xref ref-type="bibr" rid="B36">Vidal, 2017</xref>). These illusions result from the integration of spatial or temporal information of visual and auditory modalities. Previous studies have suggested the functional importance of such multisensory integration that induced more precise perception rather than unisensory perception (<xref ref-type="bibr" rid="B7">Ernst and Banks, 2002</xref>; <xref ref-type="bibr" rid="B1">Alais and Burr, 2004</xref>; <xref ref-type="bibr" rid="B5">Burr et al., 2009</xref>; <xref ref-type="bibr" rid="B12">Hartcher-O&#x2019;Brien et al., 2014</xref>; <xref ref-type="bibr" rid="B25">Murai and Yotsumoto, 2018</xref>). Investigating multisensory integration reveals how we perceive our world precisely in daily life.</p>
<p>Multimodal integration can be determined by the weighted average of the inputs of each modality (<xref ref-type="bibr" rid="B7">Ernst and Banks, 2002</xref>; <xref ref-type="bibr" rid="B1">Alais and Burr, 2004</xref>; <xref ref-type="bibr" rid="B19">Klink et al., 2011</xref>; <xref ref-type="bibr" rid="B34">Tomassini et al., 2011</xref>; <xref ref-type="bibr" rid="B29">Rohe and Noppeney, 2015a</xref>,<xref ref-type="bibr" rid="B30">b</xref>, <xref ref-type="bibr" rid="B31">2018</xref>; <xref ref-type="bibr" rid="B35">Vercillo and Gori, 2015</xref>; <xref ref-type="bibr" rid="B28">Rohe et al., 2019</xref>). When visual and auditory stimuli are presented, this information is processed separately by the visual and auditory domains. The perception of visual and auditory stimuli is represented by the relative contributions of visual and auditory information. Previous studies have suggested that the weight of this integration depends on which stimulus one attends to. <xref ref-type="bibr" rid="B36">Vidal (2017)</xref> demonstrated that when a beep and flash pair was presented with a small temporal gap, the perceived timing of this pair was captured by the attended stimulus; when participants attended to the flash, the perceived timing of this pair was shifted to the timing of the flash, and vice versa. This attentional effect on the weight of integration has been observed in various modes, such as timing or location perception (<xref ref-type="bibr" rid="B35">Vercillo and Gori, 2015</xref>; <xref ref-type="bibr" rid="B36">Vidal, 2017</xref>).</p>
<p>The effects of attention on multimodal integration have also been investigated in the domain of duration perception. Multiple studies demonstrated that when visual and auditory stimuli were presented with slightly different durations, the perceived duration was influenced by the duration of the auditory stimulus regardless of the attended modality (<xref ref-type="bibr" rid="B19">Klink et al., 2011</xref>; <xref ref-type="bibr" rid="B16">Heron et al., 2013</xref>; <xref ref-type="bibr" rid="B27">Ortega et al., 2014</xref>), except for some situation such as when the auditory stimulus is of low intensity (<xref ref-type="bibr" rid="B27">Ortega et al., 2014</xref>). These results suggest that, in most situations, the auditory modality dominates the visual modality in the duration integration, and attention does not affect the weights of each modality unless the auditory signal is weak.</p>
<p>However, since previous studies have used visual and auditory stimuli with different durations, whether auditory dominance subsists when using the same physical stimulus duration remains unclear. <xref ref-type="bibr" rid="B19">Klink et al. (2011)</xref> suggested that the perceived duration of a visual stimulus can be influenced not only by the duration but also by the onset and offset timing of the simultaneously presented irrelevant auditory stimulus. In other words, the auditory dominance reported earlier may have been a consequence of each modality&#x2019;s different onset and offset timing, which influenced the duration integration. The studies that reported auditory dominance (<xref ref-type="bibr" rid="B19">Klink et al., 2011</xref>; <xref ref-type="bibr" rid="B16">Heron et al., 2013</xref>; <xref ref-type="bibr" rid="B27">Ortega et al., 2014</xref>) have not examined this possibility.</p>
<p>In this study, we examined whether attention affects the weight of the attended modality in the auditory&#x2013;visual duration integration while controlling the physical onset&#x2013;offset timings of the stimuli. It was important that the duration of the visual and the auditory stimuli were perceptually different, while the onset and offset timings of the stimuli were the same. One solution to achieve this was to use the 10 Hz visual flicker and the 10 Hz auditory flutter: The visual flicker at around 10 Hz tends to be perceived as longer than the actual duration (<xref ref-type="bibr" rid="B18">Kanai et al., 2006</xref>; <xref ref-type="bibr" rid="B15">Herbst et al., 2013</xref>; <xref ref-type="bibr" rid="B13">Hashimoto and Yotsumoto, 2015</xref>, <xref ref-type="bibr" rid="B14">2018</xref>; <xref ref-type="bibr" rid="B37">Yuasa and Yotsumoto, 2015</xref>), while the auditory flutter at around 10 Hz tends to be perceived as shorter than the actual duration (<xref ref-type="bibr" rid="B6">Droit-Volet, 2010</xref>; <xref ref-type="bibr" rid="B37">Yuasa and Yotsumoto, 2015</xref>). Thus, using the simultaneous presentation of the 10 Hz visual flicker and auditory flutter, we can investigate the attentional effects on the perceived duration while assuring physically the same onset and offset timings but perceptually different durations.</p>
<p>In the task, we asked participants to attend only to the visual stimulus, only the auditory stimulus, or both visual and auditory stimuli. We hypothesized that the weight of the perceived duration of each stimulus was affected by the modality to which the participants attended. For example, the perceived duration of a visual flicker tends to be longer than the actual duration (<xref ref-type="bibr" rid="B37">Yuasa and Yotsumoto, 2015</xref>). Therefore, we predicted that when the participants attended to the visual flicker, the weights on the flicker would increase, resulting in the simultaneous presentation of flicker, and flutter stimuli would be perceived longer. In contrast, the perceived duration of auditory flutter tends to be shorter (<xref ref-type="bibr" rid="B37">Yuasa and Yotsumoto, 2015</xref>). Therefore, we predicted that when the participants attended to the auditory flutter, the weights on the flutter would increase, resulting in the simultaneous presentation of flicker and flutter stimuli to be perceived shorter.</p>
<p>Moreover, we applied Bayesian hierarchical modeling to represent the duration integration process. Previous studies investigated the attentional effect on duration integration by a simple comparison of the perceived duration of visual and auditory stimuli across conditions without estimating the weights of each modality (<xref ref-type="bibr" rid="B19">Klink et al., 2011</xref>; <xref ref-type="bibr" rid="B16">Heron et al., 2013</xref>; <xref ref-type="bibr" rid="B27">Ortega et al., 2014</xref>). However, this simple comparison may not be optimal for the duration-integration across modalities. The mechanism underlying duration perception has a complex hierarchical structure with a combination of modality-specific and modality-independent processing (<xref ref-type="bibr" rid="B19">Klink et al., 2011</xref>; <xref ref-type="bibr" rid="B33">Stauffer et al., 2012</xref>; <xref ref-type="bibr" rid="B16">Heron et al., 2013</xref>; <xref ref-type="bibr" rid="B24">Murai et al., 2016</xref>). Specifically, the duration information is encoded separately in each sensory modality (<xref ref-type="bibr" rid="B3">Becker and Rasmussen, 2007</xref>; <xref ref-type="bibr" rid="B24">Murai et al., 2016</xref>; <xref ref-type="bibr" rid="B23">Motala et al., 2018</xref>), and then integrated across modalities (<xref ref-type="bibr" rid="B19">Klink et al., 2011</xref>; <xref ref-type="bibr" rid="B33">Stauffer et al., 2012</xref>; <xref ref-type="bibr" rid="B16">Heron et al., 2013</xref>; <xref ref-type="bibr" rid="B12">Hartcher-O&#x2019;Brien et al., 2014</xref>; <xref ref-type="bibr" rid="B27">Ortega et al., 2014</xref>). Using Bayesian hierarchical modeling, we can define a more flexible model that matches the assumption of the hierarchical process of visual and auditory integration.</p>
<p>In the present study, we used models that assumed that the perceived duration of visual and auditory stimuli was the weighted average of each modality. We estimated the weight of the duration integration and examined whether the weight was influenced by the modality to the participants attended. In addition, attention-independent auditory dominance in duration integration can result from the higher reliability of auditory temporal perception (<xref ref-type="bibr" rid="B12">Hartcher-O&#x2019;Brien et al., 2014</xref>; <xref ref-type="bibr" rid="B36">Vidal, 2017</xref>; <xref ref-type="bibr" rid="B25">Murai and Yotsumoto, 2018</xref>). To separate the effect of reliability on duration integration from the attentional effect, we created second model assuming that the weight of the duration integration is influenced not only by the attention but also by the reliabilities of each modality. Also, to further examine the role of attention on duration integration, we created the third model assuming that the weight of duration integration is influenced only by the reliabilities of each modality, not by the attention.</p>
<p>Finally, we compared these three models to examine which model is a more likely fit to the behavioral data; in other words, we determined whether the weight for each modality can be influenced only by attention, by both the attention and the reliability of each modality, or only by the reliability.</p>
</sec>
<sec id="S2" sec-type="materials|methods">
<title>Materials and Methods</title>
<sec id="S2.SS1">
<title>Participants</title>
<p>Nineteen adults (one author and 18 na&#x00EF;ve participants, ten males and nine females) with normal or corrected-to-normal vision participated in the experiment. All participants provided written informed consent to participate in the experiment in accordance with the Declaration of Helsinki. The protocol was approved by the institutional review boards of the University of Tokyo, and all experiments were carried out in accordance with the guidelines set by the Ethics Committee of the University of Tokyo.</p>
</sec>
<sec id="S2.SS2">
<title>Apparatus</title>
<p>The visual stimuli were presented on a CRT monitor (Mitsubishi Electric RDF223H, 1024 &#x00D7; 768 pixels, 120 Hz refresh rate). The auditory stimuli were presented through a USB digital-to-analog converter Focusrite audio interface Scarlett 2i4 1st Generation and MDR-XB500 headphones at 60 dB (Sony, Japan). All stimuli were generated using MATLAB 2018b (The MathWorks Inc., Natick, MA, United States) and Psychotoolbox (<xref ref-type="bibr" rid="B4">Brainard, 1996</xref>) controlled by a Mac Pro (Apple, Cupertino, CA, United States). The viewing distance was 57.3 cm, and participants were asked to stabilize their heads on a chin rest. The experiment was conducted in a dark room.</p>
</sec>
<sec id="S2.SS3">
<title>Stimuli</title>
<p>We used a white disk (37.5 cd/m<sup>2</sup>, 5&#x00B0; diameter) as the visual target stimulus and simple tone (800 Hz, 60 dB) as the auditory target stimuli. Sound levels were calibrated using a WS1361 sound level meter (Wensn). To check the synchronization between visual and auditory stimuli, we recorded the timings of the visual and auditory stimuli using a photodiode and a Focusrite audio interface. The digital inputs from the photodiode and the Focusrite audio interface were received using UCA202 U-CONTROL (Behringer, Germany). We verified the difference in each timing using Audacity software (downloaded from <ext-link ext-link-type="uri" xlink:href="https://www.audacityteam.org/">https://www.audacityteam.org/</ext-link>). The onset and offset of every simple tone applied a ramp of 1 ms.</p>
</sec>
<sec id="S2.SS4">
<title>Procedure</title>
<p><xref ref-type="fig" rid="F1">Figure 1</xref> shows the experimental procedure for the duration discrimination task. In each trial, standard and comparison stimuli were presented sequentially. The interstimulus interval (ISI) was jittered between 640 and 960 ms. The duration of the standard stimulus was fixed at 1,000 ms. The duration of the comparison stimulus was randomly chosen from seven values (700, 800, 900, 1,000, 1,100, 1,200, or 1,300 ms) for each trial. The order of the standard and comparison stimuli was counterbalanced across the trials. After a standard and a comparison stimulus were presented, the participants compared the durations of the standard and the comparison stimuli and pressed a corresponding key to report which stimulus they perceived to be longer. The intertrial interval (ITI) was fixed at 800 ms. Two types of sessions (the unimodal and cross-modal sessions) were conducted on two separate days. The experimental procedure and the stimulus configurations were identical in the two sessions.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption><p>Experimental procedure. The participants were asked to compare the duration of a standard stimulus to that of a comparison stimulus. <bold>(A)</bold> The procedure for a unimodal session in the visual condition. <bold>(B)</bold> The procedure for a unimodal session in the auditory condition. <bold>(C)</bold> The procedure for a cross-modal session.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-15-725449-g001.tif"/>
</fig>
<p>In the unimodal session, we examined the perceived duration of the visual flicker or the auditory flutter by comparing them with the perceived duration of the stable stimulus. <xref ref-type="fig" rid="F1">Figures 1A,B</xref> show the experimental procedures for the unimodal session. The two modality conditions (visual and auditory) were tested in separate blocks. In this session, a total of 448 trials were divided into four blocks in which only visual stimuli were presented (visual condition, <xref ref-type="fig" rid="F1">Figure 1A</xref>), and four blocks in which only auditory stimuli were presented (auditory condition, <xref ref-type="fig" rid="F1">Figure 1B</xref>). Each block consisted of 56 trials. The order of the blocks was counterbalanced across participants. In the visual condition, the standard stimulus of half the trials was a visual flicker, and the standard stimulus of the other half the trials was a stable white disk. The comparison stimulus was always a stable white disk. The visual flicker consisted of a repetition of white disks for 25 ms and blanks for 75 ms (10 Hz). In the auditory condition, the standard stimulus of half the trials was an auditory flutter, and the standard stimulus of the other half the trials was a simple auditory tone. The auditory flutter consisted of a repetition of simple auditory tones for 25 ms and blanks for 75 ms.</p>
<p>In the cross-modal session, we examined the attentional effect on duration integration across the visual and the auditory modalities. There were three attention conditions [visual, auditory, and simultaneous visual&#x2013;auditory (VA) attention]. In this session, a total of 672 trials were divided into four visual attention blocks (visual attention condition), four auditory attention blocks (auditory attention condition), and four VA attention blocks (VA attention condition). Each block consisted of 56 trials. The order of the blocks was counterbalanced across participants. In all three conditions, the visual and the auditory stimuli were presented simultaneously. In half the trials, the standard stimulus was the 10 Hz visual flicker and the 10 Hz auditory flutter presented simultaneously (flicker&#x2013;flutter, <xref ref-type="fig" rid="F1">Figure 1C</xref>). In the other half trials, the standard stimulus was the stable white disk and the stable auditory tone presented simultaneously (stable&#x2013;stable). The comparison stimulus was always the stable white disk and the stable auditory tone presented simultaneously. At the beginning of each block, the participants were asked to attend to the specific stimuli. In the visual attention condition, participants were asked to attend to the visual stimuli while ignoring auditory stimuli, while in the auditory attention condition, participants were asked to attend to the auditory stimuli while ignoring visual stimuli. In the VA attention condition, participants were asked to attend to both visual and auditory stimuli simultaneously. Participants then compared the duration of the standard and the comparison stimuli that they attended to and answered which stimuli they perceived to be longer by pressing a key. At the middle and end of the block, we checked whether the participants attended to the stimuli that they were asked to attend. The instruction appeared on the screen for participants to press the left-, up-, or right-arrow key to report which modality they attended to in that block. The left-, up-, or right-arrow keys corresponded to the visual, visual, auditory, or auditory stimulus durations, respectively. If they attended to the stimuli that they were not asked, the data from that block were discarded, and the block with the same condition was appended to the end of the session.</p>
</sec>
<sec id="S2.SS5">
<title>Data Analysis</title>
<p>In each condition, we fitted a psychometric function to the data and calculated the duration of the comparison stimulus that was subjectively equivalent to the standard stimulus of 1,000 ms [i.e., the point of subjective equivalence (PSE)]. PSE indicates the duration in which the participants responded with a 50% probability that the comparison stimulus lasted longer than the standard stimulus.</p>
<p>For the unimodal session, we analyzed the Bayes factor for repeated measures analysis of variance (ANOVA) using JASP free online software (v.0.10<sup><xref ref-type="fn" rid="footnote1">1</xref></sup>, University of Amsterdam, Netherlands). We tested the hypothesis of the opposite duration distortions, where the perceived duration of the visual flicker was longer than the actual duration, while the perceived duration of the auditory flutter was shorter. The Bayes factor is the ratio of likelihood: the probability of the data given the opposite distortions between these stimuli (H<sub>1</sub>), divided by the probability of the data given no distortions (H<sub>0</sub>) (Eq. 1). Similarly, for the cross-modal session, we analyzed the Bayes factor to test the hypothesis that the perceived duration of flicker and flutter simultaneous presentation is influenced by which modality the participants attended to, the perceived duration influenced by which modality they attended to H<sub>1</sub> as opposed to no influence (H<sub>0</sub>).</p>
<disp-formula id="S2.Ex1"><label>(1)</label><mml:math id="M1" display="block">
<mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mn>10</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mo lspace="2.5pt" rspace="2.5pt" stretchy="false">|</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">H</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mo lspace="2.5pt" rspace="2.5pt" stretchy="false">|</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">H</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo rspace="5.3pt">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>The Bayes factor (<italic>BF</italic><sub><italic>10</italic></sub>) suggests the hypothesis that our results are favored. A value of <italic>BF</italic><sub><italic>10</italic></sub> larger than 1 indicates that our results favor H<sub>1</sub>, whereas a value of <italic>BF</italic><sub><italic>10</italic></sub> smaller than 1 indicates that our results favor H<sub>0</sub>. We followed the evidence categories for the Bayes factor shown in <xref ref-type="fig" rid="F2">Figure 2</xref> to interpret <italic>BF</italic><sub><italic>10</italic></sub> as the strength of evidence.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption><p>Evidence categories for the Bayes factor (<xref ref-type="bibr" rid="B17">Jeffreys, 1961</xref>).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-15-725449-g002.tif"/>
</fig>
<p>Assuming that participants perceived the duration of flicker&#x2013;flutter simultaneous presentation by integrating the perceived duration of visual flicker and auditory flutter stimulus, we analyzed how much they put weights on the perceived duration of each flicker and flutter stimulus, using the Bayesian hierarchical model (<xref ref-type="fig" rid="F3">Figure 3</xref>). We modeled this assumption using the following formula:</p>
<disp-formula id="S2.Ex2"><label>(2)</label><mml:math id="M2" display="block">
<mml:mrow>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>S</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>V</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>A</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mo>&#x223C;</mml:mo>
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>m</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>P</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>S</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>E</mml:mi>
<mml:mi>s</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>V</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>-</mml:mo>
<mml:msub>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>P</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>S</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>E</mml:mi>
<mml:mi>s</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>u</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>d</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi mathvariant="normal">&#x03C4;</mml:mi>
<mml:mi>s</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption><p>Bayesian hierarchical model with the weight by attention. <italic>A</italic> and <italic>a</italic> represent each attention condition. <italic>S</italic> and <italic>s</italic> represent each subject.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-15-725449-g003.tif"/>
</fig>
<p>where <inline-formula><mml:math id="INEQ2"><mml:mrow><mml:mi>P</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x2062;</mml:mo><mml:msubsup><mml:mi>E</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mo>,</mml:mo><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>V</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>A</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:msubsup></mml:mrow></mml:math></inline-formula> represents the PSE of flicker&#x2013;flutter stimuli in the cross-modal session, and in each attention condition <italic>a</italic> (<italic>a</italic> = 1,2,3. each corresponds to visual, auditory, VA attention condition) and each subject <italic>s</italic> (<italic>s</italic> = 1,2,&#x2026;,19. each corresponds to individual participants). <inline-formula><mml:math id="INEQ3"><mml:mrow><mml:mi>P</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x2062;</mml:mo><mml:msubsup><mml:mi>E</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>V</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>i</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>s</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:msubsup></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="INEQ4"><mml:mrow><mml:mi>P</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x2062;</mml:mo><mml:msubsup><mml:mi>E</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>u</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>d</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:msubsup></mml:mrow></mml:math></inline-formula> represent the PSE of the visual flicker or auditory flutter stimulus in the unimodal session, respectively, and in each subject <italic>s</italic>. <italic>w</italic><sub><italic>a,s</italic></sub> represents the weight of the perceived duration of flicker stimulus in each condition and subject. <inline-formula><mml:math id="INEQ5"><mml:msubsup><mml:mi mathvariant="normal">&#x03C4;</mml:mi><mml:mi>s</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:math></inline-formula> represents the noise parameter of the VA integration in each subject <italic>s</italic>. We assumed that the perceived duration of flicker&#x2013;flutter stimuli in the cross-modal session followed a normal distribution with the weighted average of the perceived durations between the visual and auditory stimuli. We assumed that the variance of this distribution was integration noise. We set a weakly informative prior to this integration noise <inline-formula><mml:math id="INEQ6"><mml:msubsup><mml:mi mathvariant="normal">&#x03C4;</mml:mi><mml:mi>s</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:math></inline-formula> as follows:</p>
<disp-formula id="S2.Ex3"><label>(3)</label><mml:math id="M3" display="block">
<mml:mrow>
<mml:mpadded width="+2.8pt">
<mml:msubsup>
<mml:mi mathvariant="normal">&#x03C4;</mml:mi>
<mml:mi>s</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mpadded>
<mml:mo rspace="5.3pt">&#x223C;</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mi>H</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mo>-</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>u</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>h</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>15</mml:mn>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>We assumed that <italic>w</italic><sub><italic>a,s</italic></sub>, the weight in each condition, and the trial should be normally distributed from the group-level weight with some individual differences. We could not construct the model that <italic>w</italic><sub><italic>a,s</italic></sub> is directly distributed from the group-level weight because the <italic>w</italic><sub><italic>a,s</italic></sub> should be between 0 and 1, while the normal distribution should be between negative infinity and positive infinity. To solve this, we first transformed <italic>w</italic><sub><italic>a,s</italic></sub> using the probit transformation as follows:</p>
<disp-formula id="S2.Ex4"><label>(4)</label><mml:math id="M4" display="block">
<mml:mrow>
<mml:mpadded width="+2.8pt">
<mml:msub>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.3pt">&#x2190;</mml:mo>
<mml:mrow>
<mml:mtext>&#x03A6;</mml:mtext>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x03B8;</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo rspace="5.3pt">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>Then we assumed that this transformed value &#x03B8;<sub><italic>a</italic>,<italic>s</italic></sub> followed the normal distribution as follows:</p>
<disp-formula id="S2.Ex5"><label>(5)</label><mml:math id="M5" display="block">
<mml:mrow>
<mml:mpadded width="+2.8pt">
<mml:msub>
<mml:mi mathvariant="normal">&#x03B8;</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.3pt">&#x223C;</mml:mo>
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>m</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mtext>&#x03A6;</mml:mtext>
<mml:mrow>
<mml:mo>-</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x03BC;</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi mathvariant="normal">&#x03C3;</mml:mi>
<mml:mi>a</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where &#x03BC;<sub><italic>a</italic></sub> represents the group-level weight in each condition and <inline-formula><mml:math id="INEQ9"><mml:msubsup><mml:mi mathvariant="normal">&#x03C3;</mml:mi><mml:mi>a</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:math></inline-formula> represents the variance of the weights across participants. We set an uninformative prior and weakly informative prior to these parameters as follows:</p>
<disp-formula id="S2.Ex6"><label>(6)</label><mml:math id="M6" display="block">
<mml:mrow>
<mml:mpadded width="+2.8pt">
<mml:msub>
<mml:mi mathvariant="normal">&#x03BC;</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.3pt">&#x2190;</mml:mo>
<mml:mrow>
<mml:mtext>Uniform</mml:mtext>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="S2.Ex7"><label>(7)</label><mml:math id="M7" display="block">
<mml:mrow>
<mml:mpadded width="+2.8pt">
<mml:msubsup>
<mml:mi mathvariant="normal">&#x03C3;</mml:mi>
<mml:mi>a</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mpadded>
<mml:mo rspace="5.3pt">&#x223C;</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mi>H</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mo>-</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>u</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>h</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>0.5</mml:mn>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>The precision of the temporal perception in the auditory modality is better than that in the visual modality (<xref ref-type="bibr" rid="B11">Guttman et al., 2008</xref>; <xref ref-type="bibr" rid="B5">Burr et al., 2009</xref>), which means higher reliability in auditory temporal perception. This difference in reliability between visual and auditory modalities can affect the weight in temporal integration (<xref ref-type="bibr" rid="B12">Hartcher-O&#x2019;Brien et al., 2014</xref>; <xref ref-type="bibr" rid="B36">Vidal, 2017</xref>; <xref ref-type="bibr" rid="B25">Murai and Yotsumoto, 2018</xref>). We created the second model combining weight by attention and reliability (<xref ref-type="fig" rid="F4">Figure 4</xref>). This model is similar to the previous model mentioned above, except for the assumption that the integration weight is influenced by attention and the reliability of each modality. We assumed weight by reliability as follows:</p>
<disp-formula id="S2.Ex8"><label>(8)</label><mml:math id="M8" display="block">
<mml:mrow>
<mml:msubsup>
<mml:mi>w</mml:mi>
<mml:mi>s</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>R</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x2190;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>D</mml:mi>
<mml:mi>s</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>V</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>-</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>D</mml:mi>
<mml:mi>s</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>V</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>-</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>D</mml:mi>
<mml:mi>s</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>u</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>d</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>-</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption><p>Bayesian hierarchical model with the weights by attention and reliability. <italic>A</italic> and <italic>a</italic> represent each attention condition. <italic>S</italic> and <italic>s</italic> represent each subject.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-15-725449-g004.tif"/>
</fig>
<p>where <inline-formula><mml:math id="INEQ10"><mml:msubsup><mml:mi>w</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>R</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> represents the weight by reliability for the perceived duration of the flicker stimulus in each condition and subject. <inline-formula><mml:math id="INEQ11"><mml:mrow><mml:mi>S</mml:mi><mml:mo>&#x2062;</mml:mo><mml:msubsup><mml:mi>D</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>V</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>i</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>s</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:msubsup></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="INEQ12"><mml:mrow><mml:mi>S</mml:mi><mml:mo>&#x2062;</mml:mo><mml:msubsup><mml:mi>D</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>u</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>d</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:msubsup></mml:mrow></mml:math></inline-formula> represent the reciprocal of slope parameters in psychometric function for visual stimulus or auditory stimulus, respectively, and in each condition and subject. If <inline-formula><mml:math id="INEQ13"><mml:mrow><mml:mi>S</mml:mi><mml:mo>&#x2062;</mml:mo><mml:msubsup><mml:mi>D</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>V</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>i</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>s</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:msubsup></mml:mrow></mml:math></inline-formula> is smaller than <inline-formula><mml:math id="INEQ14"><mml:mrow><mml:mi>S</mml:mi><mml:mo>&#x2062;</mml:mo><mml:msubsup><mml:mi>D</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>A</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>u</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>d</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:msubsup></mml:mrow></mml:math></inline-formula>, in other words, higher reliability in visual stimuli than in auditory stimuli, the weight by reliability to flicker stimulus (<inline-formula><mml:math id="INEQ15"><mml:msubsup><mml:mi>w</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>R</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula>) increases. Then, we combined this weight by reliability with weight by attention as follows:</p>
<disp-formula id="S2.Ex9"><label>(9)</label><mml:math id="M9" display="block">
<mml:mrow>
<mml:msub>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2190;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msubsup>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>A</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>w</mml:mi>
<mml:mi>s</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>R</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>A</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>w</mml:mi>
<mml:mi>s</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>R</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>-</mml:mo>
<mml:msubsup>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>A</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>-</mml:mo>
<mml:msubsup>
<mml:mi>w</mml:mi>
<mml:mi>s</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>R</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula><mml:math id="INEQ16"><mml:msubsup><mml:mi>w</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mo>,</mml:mo><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mi>A</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> represents the weight by attention for the perceived duration of the flicker stimulus in each condition and subject, as in the previous model. <italic>w</italic><sub><italic>a,s</italic></sub> represents the combined weights of attention and reliability, respectively.</p>
<p>To compare the models with and without the assumption of attentional weight, we created the third model with weight by reliability (<xref ref-type="fig" rid="F5">Figure 5</xref>). This model is also similar to the previous two models mentioned above, except for the assumption that the integration weight is influenced only by the reliabilities of each modality as in Eq. 8.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption><p>Bayesian hierarchical model with the weight by reliability. <italic>A</italic> and <italic>a</italic> represent each attention condition. <italic>S</italic> and <italic>s</italic> represent each subject.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-15-725449-g005.tif"/>
</fig>
<p>These three models were created using R (version 3.6.2) and rstan (version 2.19.3; <xref ref-type="bibr" rid="B32">Stan Development Team, 2019</xref>). We ran four chains and sampled 30,000 samples with 15,000 warm-ups within each chain. We applied a thinning parameter of 15 to reduce the effect of autocorrelation. We checked the convergence of the obtained samples using <inline-formula><mml:math id="INEQ17"><mml:mover accent="true"><mml:mi>R</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:math></inline-formula> and the effective sample size (ESS). If <inline-formula><mml:math id="INEQ18"><mml:mover accent="true"><mml:mi>R</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:math></inline-formula> is less than 1.1, and ESS is larger than 10% of the total sample size, we can interpret that there is sufficient convergence of the obtained samples.</p>
<p>We analyzed Bayes factor to compare these three models using a method of bridgesampling (<xref ref-type="bibr" rid="B9">Gronau et al., 2017</xref>) to estimate the marginal likelihood by the package of bridgesampling [version 1.0-0 (<xref ref-type="bibr" rid="B10">Gronau et al., 2020</xref>)] as follows:</p>
<disp-formula id="S2.Ex10"><label>(10)</label><mml:math id="M10" display="block">
<mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>M</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mo lspace="2.5pt" rspace="2.5pt" stretchy="false">|</mml:mo>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mo lspace="2.5pt" rspace="2.5pt" stretchy="false">|</mml:mo>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <italic>M</italic><sub><italic>1</italic></sub> represents the model assuming that weight is influenced only by attention, and <italic>M</italic><sub><italic>2</italic></sub> represents the model assuming that weight is influenced by attention and reliability. The value of <italic>BF</italic><sub><italic>M1M2</italic></sub> larger than 1 indicates that our results favor the model assuming the weight by the attention, whereas a BF value smaller than 1 means that our results favor the model assuming the weights by attention and reliability. In addition, we analyzed the Bayes factor comparing a model assuming weight by attention and reliability with a model assuming weight only by reliability and examined whether the assumption of attentional weight improved the model of duration integration.</p>
</sec>
</sec>
<sec sec-type="results" id="S3">
<title>Results</title>
<p><xref ref-type="fig" rid="F6">Figure 6</xref> shows PSEs in the unimodal session in each modality condition (visual and auditory) and each stimulus condition (stable and flicker/flutter). The PSE in the visual flicker condition was larger than that in the visual and stable condition. In comparison, the PSE in the auditory flutter condition was smaller than that in the auditory stable condition. We applied these PSEs to Bayesian repeated measures ANOVA to compare the Bayes factors between the model including the interaction effect between stimulus and modality condition, and the null model (<xref ref-type="table" rid="T1">Table 1</xref>). We found extreme evidence supporting the model, including the interaction effect between stimulus and modality condition (Bayes factor = 1.0 &#x00D7; 10<sup>3</sup>). The results suggest that visual flicker and auditory flutter induced opposite duration distortions, while the visual or auditory stable stimuli did not induce duration distortions, which was similar to that found in a previous study (<xref ref-type="bibr" rid="B37">Yuasa and Yotsumoto, 2015</xref>).</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption><p>Group results of points of subjective equivalence (PSEs) in unimodal session. The dots indicate the means of PSEs in each condition. Error bars indicate a 95% credible interval.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-15-725449-g006.tif"/>
</fig>
<table-wrap position="float" id="T1">
<label>TABLE 1</label>
<caption><p>Bayes factors of Bayesian repeated measures ANOVA in the unimodal session.</p></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left">Models</td>
<td valign="top" align="center">P(M)</td>
<td valign="top" align="center">P(M|data)</td>
<td valign="top" align="center">BF<sub>M</sub></td>
<td valign="top" align="center">BF<sub>10</sub></td>
<td valign="top" align="center">Error %</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Null model (incl. subject)</td>
<td valign="top" align="center">0.200</td>
<td valign="top" align="center">9.180e-4</td>
<td valign="top" align="center">0.004</td>
<td valign="top" align="center">1.000</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">Modality + Stimulus + Modality &#x00D7; Stimulus</td>
<td valign="top" align="center">0.200</td>
<td valign="top" align="center">0.959</td>
<td valign="top" align="center">93.277</td>
<td valign="top" align="center">1044.544</td>
<td valign="top" align="center">1.592</td>
</tr>
<tr>
<td valign="top" align="left">Modality</td>
<td valign="top" align="center">0.200</td>
<td valign="top" align="center">0.030</td>
<td valign="top" align="center">0.125</td>
<td valign="top" align="center">33.039</td>
<td valign="top" align="center">1.008</td>
</tr>
<tr>
<td valign="top" align="left">Modality + Stimulus</td>
<td valign="top" align="center">0.200</td>
<td valign="top" align="center">0.010</td>
<td valign="top" align="center">0.039</td>
<td valign="top" align="center">10.415</td>
<td valign="top" align="center">2.791</td>
</tr>
<tr>
<td valign="top" align="left">Stimulus</td>
<td valign="top" align="center">0.200</td>
<td valign="top" align="center">3.107e-4</td>
<td valign="top" align="center">0.001</td>
<td valign="top" align="center">0.338</td>
<td valign="top" align="center">11.480</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn><p><italic>All models include the subject. P(M) means the prior probability in each model. P(M|data) means the posterior probability given the obtained data. BF<sub><italic>M</italic></sub> means the ratio of posterior odds to prior odds, while BF<sub>10</sub> means the relative value of BF<sub><italic>M</italic></sub> to that of null model.</italic></p></fn>
</table-wrap-foot>
</table-wrap>
<p><xref ref-type="fig" rid="F7">Figure 7</xref> shows PSEs in the cross-modal session in each attention condition (visual, auditory, and VA) and each stimulus condition (stable&#x2013;stable and flicker&#x2013;flutter). The PSEs in the flicker&#x2013;flutter condition were smaller than those in the stable&#x2013;stable condition. In addition, the PSEs in the stable&#x2013;stable condition were almost the same across each attention condition, while in the flicker&#x2013;flutter condition, the PSE while attending to the flicker tended to be larger than that found while attending to the flutter. We applied these PSEs to Bayesian repeated measures ANOVA to compare Bayes factors between each model, including the main effect or the interaction effect between stimulus and attention condition, and the null model (<xref ref-type="table" rid="T2">Table 2</xref>). We found extreme evidence supporting the model, including the main effect of stimulus condition (Bayes factors = 7.8 &#x00D7; 10<sup>4</sup>), which indicated that simultaneous flicker and flutter presentation induced duration compression in all attention conditions. We also found extreme evidence supporting the model, including the interaction effect (Bayes factor = 1.4 &#x00D7; 10<sup>4</sup>). This interaction model suggests that paying attention to one or the other modality influences the perceived duration of the flicker and flutter simultaneous presentation, but not that of the stable stimuli.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption><p>Group results of PSEs in cross-modal session. The dots indicate the means of PSEs in each condition. Error bars indicate a 95% credible interval.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-15-725449-g007.tif"/>
</fig>
<table-wrap position="float" id="T2">
<label>TABLE 2</label>
<caption><p>Bayes factors of Bayesian repeated measures ANOVA in the cross-modal session.</p></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="justify" colspan="6">Model comparison</td>
</tr>
<tr><td colspan="6"><hr/></td></tr>
<tr>
<td valign="top" align="left">Models</td>
<td valign="top" align="center">P(M)</td>
<td valign="top" align="center">P(M|data)</td>
<td valign="top" align="center">BF<sub>M</sub></td>
<td valign="top" align="center">BF<sub>10</sub></td>
<td valign="top" align="center">Error %</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Null model (incl. subject)</td>
<td valign="top" align="center">0.200</td>
<td valign="top" align="center">8.339e-6</td>
<td valign="top" align="center">3.336e-5</td>
<td valign="top" align="center">1.000</td>
<td/>
</tr>
<tr>
<td valign="top" align="left">Stimulus</td>
<td valign="top" align="center">0.200</td>
<td valign="top" align="center">0.649</td>
<td valign="top" align="center">7.388</td>
<td valign="top" align="center">77793.649</td>
<td valign="top" align="center">0.654</td>
</tr>
<tr>
<td valign="top" align="left">Attention + Stimulus</td>
<td valign="top" align="center">0.200</td>
<td valign="top" align="center">0.228</td>
<td valign="top" align="center">1.179</td>
<td valign="top" align="center">27304.554</td>
<td valign="top" align="center">2.112</td>
</tr>
<tr>
<td valign="top" align="left">Attention + Stimulus + Attention &#x00D7; Stimulus</td>
<td valign="top" align="center">0.200</td>
<td valign="top" align="center">0.124</td>
<td valign="top" align="center">0.564</td>
<td valign="top" align="center">14812.355</td>
<td valign="top" align="center">2.440</td>
</tr>
<tr>
<td valign="top" align="left">Attention</td>
<td valign="top" align="center">0.200</td>
<td valign="top" align="center">2.026e-6</td>
<td valign="top" align="center">8.103e-6</td>
<td valign="top" align="center">0.243</td>
<td valign="top" align="center">1.079</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn><p><italic>All models include the subject. P(M) means the prior probability in each model. P(M|data) means the posterior probability given the obtained data. BF<sub><italic>M</italic></sub> means the ratio of posterior odds to prior odds, while BF<sub>10</sub> means the relative value of BF<sub><italic>M</italic></sub> to that of null model.</italic></p></fn>
</table-wrap-foot>
</table-wrap>
<p>We assumed that the duration of integration across the visual and auditory modalities was affected by attention. We modeled this assumption using the Bayesian hierarchical model for duration integration across visual and auditory modalities (<xref ref-type="fig" rid="F3">Figure 3</xref>). <xref ref-type="fig" rid="F8">Figure 8A</xref> shows the relationship between the actual PSE and predicted PSE simulated by this Bayesian hierarchical model, assuming only the weight by attention. The 95% highest-density-intervals (HDI) of almost all predicted PSEs included the dotted line, which indicated that the predicted PSE was equal to the actual PSE. Thus, this model successfully predicted actual PSE. We analyzed the posterior distributions for the group-level weights by attention (&#x03BC;<sub><italic>a</italic></sub>) in each condition (<xref ref-type="fig" rid="F8">Figure 8B</xref>). The value of the weight close to 1 meant that the duration information for the visual flicker stimulus had more weights in the VA integration, whereas the value of the weight close to 0 meant that the duration information for the auditory flutter stimulus had more weights. The modes of posterior distributions of &#x03BC;<sub><italic>a</italic></sub>in all attention conditions were smaller than 0.5, which indicated a tendency toward attention-independent auditory dominance. Although auditory dominance existed, we could also observe that the posterior distribution of &#x03BC;<sub><italic>a</italic></sub> was smaller in the auditory attention condition than in the VA and visual attention conditions.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption><p>Results of the model assuming the weight by attention. <bold>(A)</bold> The actual and predicted PSE. The black dot indicates the mode of the simulation in each subject and condition. The black line indicates 95% highest-density-interval (HDI). The dotted line indicates the predicted PSE equal to the actual PSE. <bold>(B)</bold> The posterior distributions for the group-level weights (&#x03BC;<sub><italic>a</italic></sub>) in each condition. The white dot indicates the mode of the posterior distribution. The black line indicates 95% HDI. <bold>(C)</bold> The posterior distributions for the difference of the group-level weights (&#x03BC;<sub><italic>a</italic></sub>) across attention conditions. The white dot indicates the mode of the posterior distribution. The black line indicates 95% HDI. Vis, Aud, and VA represent visual, auditory, and VA attention conditions, respectively. Comparison pairs represent the difference between each attention condition (i.e., Vis&#x2013;Aud represents the difference between visual and auditory attention conditions).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-15-725449-g008.tif"/>
</fig>
<p>To investigate whether attention to the stimulus changed the weight, we analyzed the posterior distribution for the difference in the group-level weights across attention conditions (<xref ref-type="fig" rid="F8">Figure 8C</xref>). The 95% HDI of the posterior distribution far from 0 can be interpreted as the weights being different between each attention condition pair. The posterior distribution for the difference between the visual and auditory conditions was far from 0, and the 95% HDI of the difference was above 0. This indicated that the weight of the visual information was higher when attending to the visual flicker than when attending to the auditory flutter. The 95% HDIs of other comparison pairs, the comparison with VA attention condition, included 0, which meant the weights between VA attention condition and others were not different. These results suggest that the weight of duration integration across visual and auditory modalities was changed by attention; the weight of the duration information of the attended modality increased.</p>
<p>The difference in reliability in each modality can also affect the weight of duration integration across visual and auditory stimuli. Therefore, the tendency of auditory dominance in all attention conditions may result from the higher reliability of auditory duration perception rather than visual duration perception. We modeled another Bayesian hierarchical model, which assumed that the different PSE of flicker&#x2013;flutter across the attention condition results from the change in the weight by attention and the weight by reliability (<xref ref-type="fig" rid="F4">Figure 4</xref>). By including the reliability of this model, we removed its effect from weight by attention. <xref ref-type="fig" rid="F9">Figure 9A</xref> shows the relationship between the actual PSE and the predicted PSE simulated by this Bayesian hierarchical model, assuming the weights by attention and reliability. The 95% HDIs of almost all predicted PSEs included the dotted line, which indicated that the predicted PSE was equal to the actual PSE. Thus, this model successfully predicted the actual PSE. <xref ref-type="fig" rid="F9">Figure 9B</xref> shows the posterior distributions for the group-level weights based on attention (&#x03BC;<sub><italic>a</italic></sub>).&#x03BC;<sub><italic>a</italic></sub> differed across attention conditions. However, unlike the previous model, there was no tendency for auditory dominance. This difference between models implies that auditory dominance may result from the difference in reliability between each modality and that the effect of attention on the weight remains even when the effect of reliability is controlled.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption><p>Results of the model assuming the weight by attention and reliability. <bold>(A)</bold> The actual and predicted PSE. The black dot indicates the mode of the simulation in each subject and condition. The black line indicates 95% HDI. The dotted line indicates the predicted PSE equal to the actual PSE. <bold>(B)</bold> The posterior distributions for the group-level weights (&#x03BC;<sub><italic>a</italic></sub>) in each condition. The white dot indicates the mode of the posterior distribution. The black line indicates 95% HDI. <bold>(C)</bold> The posterior distributions for the difference of the group-level weights (&#x03BC;<sub><italic>a</italic></sub>) across attention conditions. The white dot indicates the mode of the posterior distribution. The black line indicates 95% HDI. Vis, Aud, and VA represent visual, auditory, VA attention conditions, respectively. Comparison pairs represent the difference between each attention condition (i.e., Vis&#x2013;Aud represents the difference between visual and auditory attention conditions).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-15-725449-g009.tif"/>
</fig>
<p>We analyzed the posterior distribution for the difference in the group-level weights (&#x03BC;<sub><italic>a</italic></sub>) across attention conditions (<xref ref-type="fig" rid="F9">Figure 9C</xref>). Consistent with the previous model, the posterior distribution for only the difference between the visual and auditory conditions was above 0. Thus, the implication of this model was almost the same as that of the previous model, indicating that the weights were different when attending to visual flicker and when attending to auditory flutter. These estimates by the model assuming the weights by attention and reliability also suggest that attention influences the weight of duration integration across visual and auditory modalities.</p>
<p>The previous two models demonstrated that the integration weight was influenced by which modality the participants attended to. To evaluate the assumption of attentional weight on duration integration, we modeled the third Bayesian hierarchical model, which assumed that the integration weight was influenced only by the reliabilities of each modality (<xref ref-type="fig" rid="F5">Figure 5</xref>). <xref ref-type="fig" rid="F10">Figure 10</xref> shows the relationship between the actual PSE and the predicted PSE simulated by this Bayesian hierarchical model, assuming the weights by attention and reliability. The 95% HDIs of almost all predicted PSEs included the dotted line, which indicated that the predicted PSE was equal to the actual PSE. Thus, this model successfully predicted the actual PSE.</p>
<fig id="F10" position="float">
<label>FIGURE 10</label>
<caption><p>Actual and predicted PSE by the model assuming the weight by reliability. The black dot indicates the mode of the simulation in each subject and condition. The black line indicates 95% HDI. The dotted line indicates the predicted PSE equal to the actual PSE.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-15-725449-g010.tif"/>
</fig>
<p>Finally, we analyzed the Bayes factor to check which model was more likely to fit the experimental data. Comparing Model 1 assuming weight by only attention, with Model 2 assuming weight by attention and reliability, we found very strong evidence supporting Model 1 (<italic>B</italic><italic>F</italic><sub><italic>M</italic>1<italic>M</italic>2</sub> = 78.658). This evidence of model comparison suggests that our results are more likely to be represented by the integration process, assuming that the weight is influenced only by attention, not by the reliability of each modality. Also, comparing Model 2 assuming weight by attention and reliability, with Model 3 assuming weight by only reliability, we found extreme evidence supporting Model 2 (<italic>B</italic><italic>F</italic><sub><italic>M</italic>2<italic>M</italic>3</sub> = 2.6794&#x00D7;10<sup>24</sup>). This evidence suggests that the assumption including attentional weight improves our models of the duration integration process.</p>
</sec>
<sec sec-type="discussion" id="S4">
<title>Discussion</title>
<p>We investigated whether directed attention affected the weight of the attended modality in auditory&#x2013;visual duration integration. Our results showed that the perceived duration of flicker and flutter changed with attention. We then applied Bayesian hierarchical models to examine whether directed attention affects the weight of the attended modality. These models indicated that the weight on the flicker tended to be larger when the participants attended the visual flicker compared to when they attended the auditory flutter, despite the effect of the reliability in each modality.</p>
<p>Previous studies that reported the auditory dominance of duration integration used multiple stimuli that were presented for different durations (<xref ref-type="bibr" rid="B19">Klink et al., 2011</xref>; <xref ref-type="bibr" rid="B16">Heron et al., 2013</xref>; <xref ref-type="bibr" rid="B27">Ortega et al., 2014</xref>). The difference in stimulus durations led to a mismatch between the onset and offset of the stimulus presented in each modality. Hence, it is possible that the auditory dominance of duration integration partially resulted from the mismatch of the onset offset of the stimuli. To control the mismatch of the timings, we used the simultaneous presentation of a visual flicker and auditory flutter. Compared to the perceived duration of the stable stimuli, the simultaneous presentation induced a shorter perceived duration, suggesting a general dominance of the auditory modality. We further found that the perceived duration of the simultaneously presented flicker and flutter was longer when attending to the visual flicker than when attending to the auditory flutter, and when the attention was directed to both modalities, the perceived duration lay in-between them.</p>
<p>Our behavioral results were mostly consistent with <xref ref-type="bibr" rid="B37">Yuasa and Yotsumoto (2015)</xref>, where they reported time dilation with the visual flicker and time compression with the auditory flutter. However, the general auditory dominance with the simultaneously presented visual flicker and the auditory flutter was only marginally significant in their study. That might be due to the statistical power and/or to the different stimulus properties used, such as brightness, loudness, pitch, and temporal frequency (<xref ref-type="bibr" rid="B22">Matthews et al., 2011</xref>; <xref ref-type="bibr" rid="B21">Matthews and Meck, 2016</xref>).</p>
<p>The results indicate that the perceived duration of flicker and flutter was influenced by the modality the participants attended to, consistent with the attentional effect in the audio&#x2013;visual timing integration or the audio&#x2013;visual spatial location integration (<xref ref-type="bibr" rid="B35">Vercillo and Gori, 2015</xref>; <xref ref-type="bibr" rid="B36">Vidal, 2017</xref>). Therefore, our results suggest that the auditory dominance reported in previous studies (<xref ref-type="bibr" rid="B19">Klink et al., 2011</xref>; <xref ref-type="bibr" rid="B16">Heron et al., 2013</xref>; <xref ref-type="bibr" rid="B27">Ortega et al., 2014</xref>) might have been affected by the timing of the stimulus presentation.</p>
<p>We also investigated the process of multisensory duration integration and estimated the weights of the attended modality using Bayesian hierarchical modeling. We used modeling to examine how attention changes the weight of each modality in duration perception. The estimations from the models support our hypothesis that the perceived duration of the attended modality was weighted more than the non-attended modality, similar to the audio&#x2013;visual timing integration or the audio&#x2013;visual spatial location integration (<xref ref-type="bibr" rid="B35">Vercillo and Gori, 2015</xref>; <xref ref-type="bibr" rid="B36">Vidal, 2017</xref>).</p>
<p>Previous studies (<xref ref-type="bibr" rid="B12">Hartcher-O&#x2019;Brien et al., 2014</xref>; <xref ref-type="bibr" rid="B25">Murai and Yotsumoto, 2018</xref>) reported that the reliability of each sensory modality affects duration integration. However, how attention and reliability each contribute to duration integration remains unknown. Therefore, in the present study, we compared the two models; one model assumed that the weight of the duration integration is influenced only by the attention, while the other model assumes that the weight is influenced by the attention and reliability of each modality. We found that, regardless of reliability, attention influences duration integration.</p>
<p>Furthermore, we found that the estimated weights differed between models. The simple model, which did not consider the effect of reliability, revealed larger weights on the auditory flutter, independent of the attended modality. In contrast, the other model, which removed the effect of reliability, revealed no auditory dominance. These results imply that the difference in the reliability between the visual and auditory modalities may lead to the previously reported auditory dominance in duration integration (<xref ref-type="bibr" rid="B19">Klink et al., 2011</xref>; <xref ref-type="bibr" rid="B16">Heron et al., 2013</xref>; <xref ref-type="bibr" rid="B27">Ortega et al., 2014</xref>).</p>
<p>Overall, the model comparison favored the models assuming the attentional effect on duration integration over the model without assuming it. The result of the model comparison supports the influence of attention in duration integration across modalities. On the other hand, the comparison also favored the model that did consider the attention, but not reliability, over the model that considered both attention and reliability. That result was unexpected, given that reliability has been reported to influence duration integration (<xref ref-type="bibr" rid="B12">Hartcher-O&#x2019;Brien et al., 2014</xref>; <xref ref-type="bibr" rid="B25">Murai and Yotsumoto, 2018</xref>). While all our models reasonably described the data, our model might not have grasped the full picture of the complexity of duration integration. For example, we did not consider the interaction between attention and reliability, while <xref ref-type="bibr" rid="B35">Vercillo and Gori (2015)</xref> demonstrated that attending to one modality affected the reliability of the attended modality. In addition, our model assumed the equal contribution of both attention and reliability to the duration integration, although attention may influence the integration to a greater extent than reliability, or vice versa. The purpose of the present study was not to find the best model to capture the complexity of the duration integration; rather, it was to investigate whether attention affects modality weight on duration integration while considering the influence of reliability. Thus, the present study serves as a starting point for future studies to examine more detailed processes.</p>
<p>Future studies can also extend our models to consider other factors that influence duration integration. For example, we could examine the integration with other senses such as tactile (<xref ref-type="bibr" rid="B34">Tomassini et al., 2011</xref>; <xref ref-type="bibr" rid="B35">Vercillo and Gori, 2015</xref>). We can also examine the influence of spatial information on integration. Temporal integration between audiovisuals becomes more difficult as the distance between each sensory input increases (<xref ref-type="bibr" rid="B8">Godfroy et al., 2003</xref>; <xref ref-type="bibr" rid="B20">Lewald and Guski, 2003</xref>). In addition, the perceived duration of the target stimulus is affected by the location of the distractor stimulus (<xref ref-type="bibr" rid="B26">Okajima and Yotsumoto, 2016</xref>). Based on the influences of spatial information on temporal integration, the attentional effect on multisensory duration integration may also be influenced by the spatial location of each stimulus.</p>
<p>In the present study, we used the 10 Hz visual flicker and auditory flutter to examine the effect of modality-specific attention on duration integration while controlling the confounding effect of the onset&#x2013;offset timing of stimuli. The behavioral results demonstrated that the perceived duration of flicker and flutter was influenced by which modality the participants attended. Moreover, our modeling results supported the idea that the modality-specific attention modulated the weight of duration integration across visual and auditory modality, regardless of the reliability of each modality. Based on the effects of attention and reliability explored by our study, further studies may: (1) explore the complex influence of attention and reliability on duration integration, and (2) add other sensory and spatial information to our models.</p>
</sec>
<sec sec-type="data-availability" id="S5">
<title>Data Availability Statement</title>
<p>The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found below: <ext-link ext-link-type="uri" xlink:href="https://osf.io/v3fqr/?view_only=b6b3ba978600405a949bbbd929ebbfcc">https://osf.io/v3fqr/?view_only=b6b3ba978600405a949bbbd929ebbfcc</ext-link>.</p>
</sec>
<sec id="S6">
<title>Ethics Statement</title>
<p>The studies involving human participants were reviewed and approved by Internal Review Board of the University of Tokyo. The patients/participants provided their written informed consent to participate in this study.</p>
</sec>
<sec id="S7">
<title>Author Contributions</title>
<p>HY and YY designed the experiments and wrote the manuscript. HY conducted the experiments and analyzed data. Both authors contributed to the article and approved the submitted version.</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="S8">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
</body>
<back>
<sec sec-type="funding-information" id="S9">
<title>Funding</title>
<p>This work was supported by Japan Society for the Promotion of Sciences (JSPS <ext-link ext-link-type="uri" xlink:href="https://www.jsps.go.jp/english/index.html">https://www.jsps.go.jp/english/index.html</ext-link>) for HY, and KAKENHI (Grant #19H01771, #19J22514) and U-Tokyo CiSHuB for YY.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alais</surname> <given-names>D.</given-names></name> <name><surname>Burr</surname> <given-names>D.</given-names></name></person-group> (<year>2004</year>). <article-title>The ventriloquist effect results from near-optimal bimodal integration.</article-title> <source><italic>Curr. Biol.</italic></source> <volume>14</volume> <fpage>257</fpage>&#x2013;<lpage>262</lpage>. <pub-id pub-id-type="doi">10.1016/j.cub.2004.01.029</pub-id> <pub-id pub-id-type="pmid">14761661</pub-id></citation></ref>
<ref id="B2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Battaglia</surname> <given-names>P. W.</given-names></name> <name><surname>Jacobs</surname> <given-names>R. A.</given-names></name> <name><surname>Aslin</surname> <given-names>R. N.</given-names></name></person-group> (<year>2003</year>). <article-title>Bayesian integration of visual and auditory signals for spatial localization.</article-title> <source><italic>J. Opt. Soc. Am. A</italic></source> <volume>20</volume>:<fpage>1391</fpage>. <pub-id pub-id-type="doi">10.1364/josaa.20.001391</pub-id> <pub-id pub-id-type="pmid">12868643</pub-id></citation></ref>
<ref id="B3"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Becker</surname> <given-names>M. W.</given-names></name> <name><surname>Rasmussen</surname> <given-names>I. P.</given-names></name></person-group> (<year>2007</year>). <article-title>The rhythm aftereffect: support for time sensitive neurons with broad overlapping tuning curves.</article-title> <source><italic>Brain Cogn.</italic></source> <volume>64</volume> <fpage>274</fpage>&#x2013;<lpage>281</lpage>. <pub-id pub-id-type="doi">10.1016/j.bandc.2007.03.009</pub-id> <pub-id pub-id-type="pmid">17478023</pub-id></citation></ref>
<ref id="B4"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brainard</surname> <given-names>D. H.</given-names></name></person-group> (<year>1996</year>). <article-title>The psychophysics toolbox.</article-title> <source><italic>Spat. Vis.</italic></source> <volume>10</volume> <fpage>433</fpage>&#x2013;<lpage>436</lpage>.</citation></ref>
<ref id="B5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Burr</surname> <given-names>D.</given-names></name> <name><surname>Banks</surname> <given-names>M. S.</given-names></name> <name><surname>Morrone</surname> <given-names>M. C.</given-names></name></person-group> (<year>2009</year>). <article-title>Auditory dominance over vision in the perception of interval duration.</article-title> <source><italic>Exp. Brain Res.</italic></source> <volume>198</volume> <fpage>49</fpage>&#x2013;<lpage>57</lpage>. <pub-id pub-id-type="doi">10.1007/s00221-009-1933-z</pub-id> <pub-id pub-id-type="pmid">19597804</pub-id></citation></ref>
<ref id="B6"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Droit-Volet</surname> <given-names>S.</given-names></name></person-group> (<year>2010</year>). <article-title>Speeding up a master clock common to time, number and length?</article-title> <source><italic>Behav. Process.</italic></source> <volume>85</volume> <fpage>126</fpage>&#x2013;<lpage>134</lpage>. <pub-id pub-id-type="doi">10.1016/j.beproc.2010.06.017</pub-id> <pub-id pub-id-type="pmid">20600693</pub-id></citation></ref>
<ref id="B7"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ernst</surname> <given-names>M. O.</given-names></name> <name><surname>Banks</surname> <given-names>M. S.</given-names></name></person-group> (<year>2002</year>). <article-title>Humans integrate visual and haptic information in a.</article-title> <source><italic>Nature</italic></source> <volume>415</volume> <fpage>429</fpage>&#x2013;<lpage>433</lpage>.</citation></ref>
<ref id="B8"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Godfroy</surname> <given-names>M.</given-names></name> <name><surname>Roumes</surname> <given-names>C.</given-names></name> <name><surname>Dauchy</surname> <given-names>P.</given-names></name></person-group> (<year>2003</year>). <article-title>Spatial variations of visual &#x2013; auditory fusion areas.</article-title> <source><italic>Perception</italic></source> <volume>32</volume> <fpage>1233</fpage>&#x2013;<lpage>1245</lpage>. <pub-id pub-id-type="doi">10.1068/p3344</pub-id> <pub-id pub-id-type="pmid">14700258</pub-id></citation></ref>
<ref id="B9"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gronau</surname> <given-names>Q. F.</given-names></name> <name><surname>Sarafoglou</surname> <given-names>A.</given-names></name> <name><surname>Matzke</surname> <given-names>D.</given-names></name> <name><surname>Ly</surname> <given-names>A.</given-names></name> <name><surname>Boehm</surname> <given-names>U.</given-names></name> <name><surname>Marsman</surname> <given-names>M.</given-names></name><etal/></person-group> (<year>2017</year>). <article-title>A tutorial on bridge sampling.</article-title> <source><italic>J. Math. Psychol.</italic></source> <volume>81</volume> <fpage>80</fpage>&#x2013;<lpage>97</lpage>. <pub-id pub-id-type="doi">10.1016/j.jmp.2017.09.005</pub-id> <pub-id pub-id-type="pmid">29200501</pub-id></citation></ref>
<ref id="B10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gronau</surname> <given-names>Q. F.</given-names></name> <name><surname>Singmann</surname> <given-names>H.</given-names></name> <name><surname>Wagenmakers</surname> <given-names>E. J.</given-names></name></person-group> (<year>2020</year>). <article-title>Bridgesampling: an R package for estimating normalizing constants.</article-title> <source><italic>J. Stat. Softw.</italic></source> <volume>92</volume> <fpage>1</fpage>&#x2013;<lpage>29</lpage>. <pub-id pub-id-type="doi">10.18637/jss.v092.i10</pub-id></citation></ref>
<ref id="B11"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Guttman</surname> <given-names>S. E.</given-names></name> <name><surname>Gilroy</surname> <given-names>L. A.</given-names></name> <name><surname>Blake</surname> <given-names>R.</given-names></name></person-group> (<year>2008</year>). <article-title>Hearing what the eyes see: auditory encoding of visual temporal sequences sharon.</article-title> <source><italic>Psychol. Sci.</italic></source> <volume>23</volume> <fpage>1</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1038/jid.2014.371</pub-id> <pub-id pub-id-type="pmid">25178106</pub-id></citation></ref>
<ref id="B12"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hartcher-O&#x2019;Brien</surname> <given-names>J.</given-names></name> <name><surname>Di Luca</surname> <given-names>M.</given-names></name> <name><surname>Ernst</surname> <given-names>M. O.</given-names></name></person-group> (<year>2014</year>). <article-title>The duration of uncertain times: audiovisual information about intervals is integrated in a statistically optimal fashion.</article-title> <source><italic>PLoS One</italic></source> <volume>9</volume>:<fpage>e89339</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0089339</pub-id> <pub-id pub-id-type="pmid">24594578</pub-id></citation></ref>
<ref id="B13"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hashimoto</surname> <given-names>Y.</given-names></name> <name><surname>Yotsumoto</surname> <given-names>Y.</given-names></name></person-group> (<year>2015</year>). <article-title>Effect of temporal frequency spectra of flicker on time perception: behavioral testing and simulations using a striatal beat frequency model.</article-title> <source><italic>Timing Time Percept.</italic></source> <volume>3</volume> <fpage>201</fpage>&#x2013;<lpage>222</lpage>. <pub-id pub-id-type="doi">10.1163/22134468-03002049</pub-id></citation></ref>
<ref id="B14"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hashimoto</surname> <given-names>Y.</given-names></name> <name><surname>Yotsumoto</surname> <given-names>Y.</given-names></name></person-group> (<year>2018</year>). <article-title>The amount of time dilation for visual flickers corresponds to the amount of neural entrainments measured by EEG.</article-title> <source><italic>Front. Comput. Neurosci.</italic></source> <volume>12</volume>:<fpage>30</fpage>. <pub-id pub-id-type="doi">10.3389/fncom.2018.00030</pub-id> <pub-id pub-id-type="pmid">29867423</pub-id></citation></ref>
<ref id="B15"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Herbst</surname> <given-names>S. K.</given-names></name> <name><surname>Javadi</surname> <given-names>A. H.</given-names></name> <name><surname>van der Meer</surname> <given-names>E.</given-names></name> <name><surname>Busch</surname> <given-names>N. A.</given-names></name></person-group> (<year>2013</year>). <article-title>How long depends on how fast-perceived flicker dilates subjective duration.</article-title> <source><italic>PLoS One</italic></source> <volume>8</volume>:<fpage>e76074</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0076074</pub-id> <pub-id pub-id-type="pmid">24194829</pub-id></citation></ref>
<ref id="B16"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Heron</surname> <given-names>J.</given-names></name> <name><surname>Hotchkiss</surname> <given-names>J.</given-names></name> <name><surname>Aaen-Stockdale</surname> <given-names>C.</given-names></name> <name><surname>Roach</surname> <given-names>N. W.</given-names></name> <name><surname>Whitaker</surname> <given-names>D.</given-names></name></person-group> (<year>2013</year>). <article-title>A neural hierarchy for illusions of time: duration adaptation precedes multisensory integration.</article-title> <source><italic>J. Vis.</italic></source> <volume>13</volume>:<fpage>4</fpage>. <pub-id pub-id-type="doi">10.1167/13.14.4</pub-id></citation></ref>
<ref id="B17"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jeffreys</surname> <given-names>H.</given-names></name></person-group> (<year>1961</year>). <source><italic>The Theory of Probability</italic></source>, <edition>3rd Edn</edition>. <publisher-loc>Oxford</publisher-loc>: <publisher-name>Oxford University Press</publisher-name>.</citation></ref>
<ref id="B18"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kanai</surname> <given-names>R.</given-names></name> <name><surname>Paffen</surname> <given-names>C. L. E.</given-names></name> <name><surname>Hogendoorn</surname> <given-names>H.</given-names></name> <name><surname>Verstraten</surname> <given-names>F. A. J.</given-names></name></person-group> (<year>2006</year>). <article-title>Time dilation in dynamic visual display.</article-title> <source><italic>J. Vis.</italic></source> <volume>6</volume> <fpage>1421</fpage>&#x2013;<lpage>1430</lpage>. <pub-id pub-id-type="doi">10.1167/6.12.8</pub-id></citation></ref>
<ref id="B19"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Klink</surname> <given-names>P. C.</given-names></name> <name><surname>Montijn</surname> <given-names>J. S.</given-names></name> <name><surname>van Wezel</surname> <given-names>R. J. A.</given-names></name></person-group> (<year>2011</year>). <article-title>Crossmodal duration perception involves perceptual grouping, temporal ventriloquism, and variable internal clock rates.</article-title> <source><italic>Atten. Percept. Psychophys.</italic></source> <volume>73</volume> <fpage>219</fpage>&#x2013;<lpage>236</lpage>. <pub-id pub-id-type="doi">10.3758/s13414-010-0010-9</pub-id> <pub-id pub-id-type="pmid">21258921</pub-id></citation></ref>
<ref id="B20"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lewald</surname> <given-names>J.</given-names></name> <name><surname>Guski</surname> <given-names>R.</given-names></name></person-group> (<year>2003</year>). <article-title>Cross-modal perceptual integration of spatially and temporally disparate auditory and visual stimuli.</article-title> <source><italic>Cogn. Brain Res.</italic></source> <volume>16</volume> <fpage>468</fpage>&#x2013;<lpage>478</lpage>. <pub-id pub-id-type="doi">10.1016/S0926-6410(03)00074-0</pub-id></citation></ref>
<ref id="B21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Matthews</surname> <given-names>W. J.</given-names></name> <name><surname>Meck</surname> <given-names>W. H.</given-names></name></person-group> (<year>2016</year>). <article-title>Temporal cognition: connecting subjective time to perception, attention, and memory.</article-title> <source><italic>Psychol. Bull.</italic></source> <volume>142</volume> <fpage>865</fpage>&#x2013;<lpage>907</lpage>. <pub-id pub-id-type="doi">10.1037/bul0000045</pub-id> <pub-id pub-id-type="pmid">27196725</pub-id></citation></ref>
<ref id="B22"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Matthews</surname> <given-names>W. J.</given-names></name> <name><surname>Stewart</surname> <given-names>N.</given-names></name> <name><surname>Wearden</surname> <given-names>J. H.</given-names></name></person-group> (<year>2011</year>). <article-title>Stimulus intensity and the perception of duration.</article-title> <source><italic>J. Exp. Psychol. Hum. Percept. Perform.</italic></source> <volume>37</volume> <fpage>303</fpage>&#x2013;<lpage>313</lpage>. <pub-id pub-id-type="doi">10.1037/a0019961</pub-id> <pub-id pub-id-type="pmid">20731508</pub-id></citation></ref>
<ref id="B23"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Motala</surname> <given-names>A.</given-names></name> <name><surname>Heron</surname> <given-names>J.</given-names></name> <name><surname>McGraw</surname> <given-names>P. V.</given-names></name> <name><surname>Roach</surname> <given-names>N. W.</given-names></name> <name><surname>Whitaker</surname> <given-names>D.</given-names></name></person-group> (<year>2018</year>). <article-title>Rate after-effects fail to transfer cross-modally: evidence for distributed sensory timing mechanisms.</article-title> <source><italic>Sci. Rep.</italic></source> <volume>8</volume>:<fpage>924</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-018-19218-z</pub-id> <pub-id pub-id-type="pmid">29343859</pub-id></citation></ref>
<ref id="B24"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Murai</surname> <given-names>Y.</given-names></name> <name><surname>Whitaker</surname> <given-names>D.</given-names></name> <name><surname>Yotsumoto</surname> <given-names>Y.</given-names></name></person-group> (<year>2016</year>). <article-title>The centralized and distributed nature of adaptation-induced misjudgments of time adaptation-induced misjudgments of time.</article-title> <source><italic>Curr. Opin. Behav. Sci.</italic></source> <volume>8</volume> <fpage>117</fpage>&#x2013;<lpage>123</lpage>. <pub-id pub-id-type="doi">10.1016/j.cobeha.2016.02.011</pub-id></citation></ref>
<ref id="B25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Murai</surname> <given-names>Y.</given-names></name> <name><surname>Yotsumoto</surname> <given-names>Y.</given-names></name></person-group> (<year>2018</year>). <article-title>Optimal multisensory integration leads to optimal time estimation.</article-title> <source><italic>Sci. Rep.</italic></source> <volume>8</volume>:<fpage>13068</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-018-31468-5</pub-id> <pub-id pub-id-type="pmid">30166608</pub-id></citation></ref>
<ref id="B26"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Okajima</surname> <given-names>M.</given-names></name> <name><surname>Yotsumoto</surname> <given-names>Y.</given-names></name></person-group> (<year>2016</year>). <article-title>Flickering task &#x2013; irrelevant distractors induce dilation of target duration depending upon cortical distance.</article-title> <source><italic>Sci. Rep.</italic></source> <volume>6</volume>:<fpage>32432</fpage>. <pub-id pub-id-type="doi">10.1038/srep32432</pub-id> <pub-id pub-id-type="pmid">27577614</pub-id></citation></ref>
<ref id="B27"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ortega</surname> <given-names>L.</given-names></name> <name><surname>Guzman-Martinez</surname> <given-names>E.</given-names></name> <name><surname>Grabowecky</surname> <given-names>M.</given-names></name> <name><surname>Suzuki</surname> <given-names>S.</given-names></name></person-group> (<year>2014</year>). <article-title>Audition dominates vision in duration perception irrespective of salience, attention, and temporal discriminability.</article-title> <source><italic>Atten. Percept. Psychophys.</italic></source> <volume>23</volume> <fpage>1</fpage>&#x2013;<lpage>7</lpage>.</citation></ref>
<ref id="B28"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rohe</surname> <given-names>T.</given-names></name> <name><surname>Ehlis</surname> <given-names>A. C.</given-names></name> <name><surname>Noppeney</surname> <given-names>U.</given-names></name></person-group> (<year>2019</year>). <article-title>The neural dynamics of hierarchical Bayesian causal inference in multisensory perception.</article-title> <source><italic>Nat. Commun.</italic></source> <volume>10</volume>:<fpage>1907</fpage>. <pub-id pub-id-type="doi">10.1038/s41467-019-09664-2</pub-id> <pub-id pub-id-type="pmid">31015423</pub-id></citation></ref>
<ref id="B29"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rohe</surname> <given-names>T.</given-names></name> <name><surname>Noppeney</surname> <given-names>U.</given-names></name></person-group> (<year>2015a</year>). <article-title>Cortical hierarchies perform bayesian causal inference in multisensory perception.</article-title> <source><italic>PLoS Biol.</italic></source> <volume>13</volume>:<fpage>e1002073</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pbio.1002073</pub-id> <pub-id pub-id-type="pmid">25710328</pub-id></citation></ref>
<ref id="B30"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rohe</surname> <given-names>T.</given-names></name> <name><surname>Noppeney</surname> <given-names>U.</given-names></name></person-group> (<year>2015b</year>). <article-title>Sensory reliability shapes perceptual inference via two mechanisms.</article-title> <source><italic>J. Vis.</italic></source> <volume>15</volume>:<fpage>22</fpage>. <pub-id pub-id-type="doi">10.1167/15.5.22</pub-id></citation></ref>
<ref id="B31"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rohe</surname> <given-names>T.</given-names></name> <name><surname>Noppeney</surname> <given-names>U.</given-names></name></person-group> (<year>2018</year>). <article-title>Reliability-weighted integration of audiovisual signals can be modulated by top-down attention.</article-title> <source><italic>eNeuro</italic></source> <volume>5</volume>:<comment>ENEURO.0315-17.2018</comment>. <pub-id pub-id-type="doi">10.1523/ENEURO.0315-17.2018</pub-id> <pub-id pub-id-type="pmid">29527567</pub-id></citation></ref>
<ref id="B32"><citation citation-type="journal"><collab>Stan Development Team</collab> (<year>2019</year>). <source><italic>RStan: The R Interface to Stan, Version 2.19.3</italic></source>. <comment>Available online at:</comment> <ext-link ext-link-type="uri" xlink:href="http://mc-stan.org/">http://mc-stan.org/</ext-link> <comment>(accessed March 18, 2020)</comment>.</citation></ref>
<ref id="B33"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stauffer</surname> <given-names>C. C.</given-names></name> <name><surname>Haldemann</surname> <given-names>J.</given-names></name> <name><surname>Troche</surname> <given-names>S. J.</given-names></name> <name><surname>Rammsayer</surname> <given-names>T. H.</given-names></name></person-group> (<year>2012</year>). <article-title>Auditory and visual temporal sensitivity: evidence for a hierarchical structure of modality-specific and modality-independent levels of temporal information processing.</article-title> <source><italic>Psychol. Res.</italic></source> <volume>76</volume> <fpage>20</fpage>&#x2013;<lpage>31</lpage>. <pub-id pub-id-type="doi">10.1007/s00426-011-0333-8</pub-id> <pub-id pub-id-type="pmid">21461936</pub-id></citation></ref>
<ref id="B34"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tomassini</surname> <given-names>A.</given-names></name> <name><surname>Gori</surname> <given-names>M.</given-names></name> <name><surname>Burr</surname> <given-names>D.</given-names></name> <name><surname>Sandini</surname> <given-names>G.</given-names></name> <name><surname>Morrone</surname> <given-names>M. C.</given-names></name></person-group> (<year>2011</year>). <article-title>Perceived duration of visual and tactile stimuli depends on perceived speed.</article-title> <source><italic>Front. Integr. Neurosci.</italic></source> <volume>5</volume>:<fpage>51</fpage>. <pub-id pub-id-type="doi">10.3389/fnint.2011.00051</pub-id> <pub-id pub-id-type="pmid">21941471</pub-id></citation></ref>
<ref id="B35"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vercillo</surname> <given-names>T.</given-names></name> <name><surname>Gori</surname> <given-names>M.</given-names></name></person-group> (<year>2015</year>). <article-title>Attention to sound improves auditory reliability in audio-tactile spatial optimal integration.</article-title> <source><italic>Front. Integr. Neurosci.</italic></source> <volume>9</volume>:<fpage>34</fpage>. <pub-id pub-id-type="doi">10.3389/fnint.2015.00034</pub-id> <pub-id pub-id-type="pmid">25999825</pub-id></citation></ref>
<ref id="B36"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vidal</surname> <given-names>M.</given-names></name></person-group> (<year>2017</year>). <article-title>Hearing flashes and seeing beeps: timing audiovisual events.</article-title> <source><italic>PLoS One</italic></source> <volume>12</volume>:<fpage>e0172028</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0172028</pub-id> <pub-id pub-id-type="pmid">28207786</pub-id></citation></ref>
<ref id="B37"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yuasa</surname> <given-names>K.</given-names></name> <name><surname>Yotsumoto</surname> <given-names>Y.</given-names></name></person-group> (<year>2015</year>). <article-title>Opposite distortions in interval timing perception for visual and auditory stimuli with temporal modulations.</article-title> <source><italic>PLoS One</italic></source> <volume>10</volume>:<fpage>e0135646</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0135646</pub-id> <pub-id pub-id-type="pmid">26292285</pub-id></citation></ref>
</ref-list>
<fn-group>
<fn id="footnote1">
<label>1</label>
<p><ext-link ext-link-type="uri" xlink:href="https://jasp-stats.org/">https://jasp-stats.org/</ext-link></p></fn>
</fn-group>
</back>
</article>