<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Hum. Neurosci.</journal-id>
<journal-title>Frontiers in Human Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Hum. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-5161</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnhum.2021.612345</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Human Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Asymmetries in Accessing Vowel Representations Are Driven by Phonological and Acoustic Properties: Neural and Behavioral Evidence From Natural German Minimal Pairs</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Riedinger</surname> <given-names>Miriam</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/894871/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Nagels</surname> <given-names>Arne</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/84587/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Werth</surname> <given-names>Alexander</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/139403/overview"/>
</contrib> 
<contrib contrib-type="author">
<name><surname>Scharinger</surname> <given-names>Mathias</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/51552/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Department of English and Linguistics, Johannes Gutenberg University</institution>, <addr-line>Mainz</addr-line>, <country>Germany</country></aff>
<aff id="aff2"><sup>2</sup><institution>Institute for German Linguistics, Philipps University</institution>, <addr-line>Marburg</addr-line>, <country>Germany</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Language and Literature, Max Planck Institute for Empirical Aesthetics</institution>, <addr-line>Frankfurt</addr-line>, <country>Germany</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Yang Zhang, University of Minnesota Health Twin Cities, United States</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Linda Polka, McGill University, Canada; Miwako Hisagi, California State University, United States</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Miriam Riedinger <email>miriedin&#x00040;uni-mainz.de</email></corresp>
<fn fn-type="other" id="fn001"><p><bold>Specialty section</bold>: This article was submitted to Speech and Language, a section of the journal Frontiers in Human Neuroscience</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>18</day>
<month>02</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>15</volume>
<elocation-id>612345</elocation-id>
<history>
<date date-type="received">
<day>30</day>
<month>09</month>
<year>2020</year>
</date>
<date date-type="accepted">
<day>26</day>
<month>01</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2021 Riedinger, Nagels, Werth and Scharinger.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Riedinger, Nagels, Werth and Scharinger</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract><p>In vowel discrimination, commonly found discrimination patterns are directional asymmetries where discrimination is faster (or easier) if differing vowels are presented in a certain sequence compared to the reversed sequence. Different models of speech sound processing try to account for these asymmetries based on either phonetic or phonological properties. In this study, we tested and compared two of those often-discussed models, namely the Featurally Underspecified Lexicon (FUL) model (Lahiri and Reetz, <xref ref-type="bibr" rid="B36">2002</xref>) and the Natural Referent Vowel (NRV) framework (Polka and Bohn, <xref ref-type="bibr" rid="B64">2011</xref>). While most studies presented isolated vowels, we investigated a large stimulus set of German vowels in a more naturalistic setting within minimal pairs. We conducted an mismatch negativity (MMN) study in a passive and a reaction time study in an active oddball paradigm. In both data sets, we found directional asymmetries that can be explained by either phonological or phonetic theories. While behaviorally, the vowel discrimination was based on phonological properties, both tested models failed to explain the found neural patterns comprehensively. Therefore, we additionally examined the influence of a variety of articulatory, acoustical, and lexical factors (e.g., formant structure, intensity, duration, and frequency of occurrence) but also the influence of factors beyond the well-known (perceived loudness of vowels, degree of openness) in depth <italic>via</italic> multiple regression analyses. The analyses revealed that the perceptual factor of perceived loudness has a greater impact than considered in the literature and should be taken stronger into consideration when analyzing preattentive natural vowel processing.</p></abstract>
<kwd-group>
<kwd>vowel discrimination</kwd>
<kwd>mismatch negativity (MMN)</kwd>
<kwd>reaction time (RT)</kwd>
<kwd>multiple regression analysis</kwd>
<kwd>perceived loudness</kwd>
</kwd-group>
<contract-sponsor id="cn001">Hessisches Ministerium f&#x000FC;r Wissenschaft und Kunst<named-content content-type="fundref-id">10.13039/501100003495</named-content></contract-sponsor>
<contract-sponsor id="cn002">Johannes Gutenberg-Universit&#x000E4;t Mainz<named-content content-type="fundref-id">10.13039/501100004033</named-content></contract-sponsor>
<counts>
<fig-count count="7"/>
<table-count count="5"/>
<equation-count count="1"/>
<ref-count count="105"/>
<page-count count="20"/>
<word-count count="15324"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="introduction" id="s1">
<title>Introduction</title>
<p>In recent years, much research has been done on the mental representations of vowels and on investigating which properties are involved in vowel discrimination. This article investigates the mental representations of vowels and compares two models that both make specific hypotheses regarding sound discrimination and mental representations of speech sounds, namely the Featurally Underspecified Lexicon (FUL) model (Lahiri and Reetz, <xref ref-type="bibr" rid="B36">2002</xref>, <xref ref-type="bibr" rid="B37">2010</xref>) and the Natural Referent Vowel (NRV) framework (Polka and Bohn, <xref ref-type="bibr" rid="B64">2011</xref>). Based on the notions that spoken language has a sequential, serial structure and those earlier events that precede later events influence the recognition or discrimination of those later events, both models have in common that they are predicting directional asymmetries in the discrimination of speech sounds: discrimination of two speech sounds is easier in one direction than in the other; therefore, it matters which sound is presented first. For example, when testing the discrimination of two vowels (e.g., [i] and [e]), one can present the vowels in two possible orders: the high vowel followed by the mid vowel ([i]&#x02014;[e]) or in the reverse order ([e]&#x02014;[i]). Both models assume that vowel discrimination is based on the nature of the mental representation and predict facilitated discrimination in one direction, but predictions about the easier presentation order are often competing. Furthermore, what separates the models are the substantially different assumptions about the features involved in discrimination processes and therefore in mental representations.</p>
<p>Within the FUL model, Lahiri and Reetz (<xref ref-type="bibr" rid="B36">2002</xref>, <xref ref-type="bibr" rid="B37">2010</xref>) made a proposition for speech perception and lexical access suggesting that speech sounds can be described with the help of abstract and underspecified feature specifications (e.g., [HIGH] for high or close vowels, such as [i]). Importantly, they also describe sound processing based on those features. Crucially, this model assumes that there can be a discrepancy between the features contained in the signal and those stored in the mental lexicon, since mental representations may be underspecified and therefore do not contain all possible features. These assumptions of underspecified mental representations express both similarities and differences to other approaches of underspecification. In common with other underspecification theories, the underspecified sound descriptions are based on the notion of <italic>minimalism</italic>. In this respect, it is postulated that only a distinct set of sound descriptors are necessary for underlying representations. But in contrast to theories like Radical Underspecification (Archangeli, <xref ref-type="bibr" rid="B5">1988</xref>) the underspecification approach in FUL is not only a theoretical means to describe certain linguistic phenomena (e.g., assimilation) but also constitutes mental representations of speech. Therefore underspecification is directly involved in speech perception and production. Additionally, in FUL sounds can be described solely with monovalent features. For example, in FUL it is believed that coronal segments (i.e., front vowels) are underspecified for a place of articulation information ([&#x02013;]) in the mental representation, but the feature [COR] can be retrieved from the auditory signal. This underspecification approach, together with the specific proposed ternary mapping process, is the reason for the resulting directional asymmetries in sound discrimination. This mapping process includes a comparison of the features obtained from the signal with those stored in the mental lexicon. Due to the underspecification of redundant features, there are three possible outcomes: a match occurs if the feature extracted from the signal has the same equivalent feature in the mental lexicon (e.g., [DOR]&#x02014;[DOR]: [u]&#x02014;[o]). A mismatch occurs if the feature taken from the signal and the feature in the underlying representation are complementary and exclude each other (e.g., [HIGH]&#x02014;[LOW]: [i]&#x02014;[a]). Last but not least, a no-mismatch occurs if a feature extracted from the signal neither mismatches with a feature of the mental lexicon nor matches it. The last setup of the mapping process is crucial in the elicitation of directional asymmetries. For example, if [COR] is extracted from the signal, this feature produces a mismatch with [DOR] in the lexicon, but if [DOR] is extracted from the signal, the result is a no-mismatch due to the underspecification of the coronal place of articulation ([&#x02013;]). These different results should become apparent when the discrimination of two vowels is tested in both possible presentation orders ([i]&#x02014;[u] vs. [u]&#x02014;[i]).</p>
<p>Several studies have shown that the presentation order with a mismatch as the result of the mapping process usually elicits larger effects than vice versa. Eulitz and Lahiri (<xref ref-type="bibr" rid="B20">2004</xref>) conducted an ERP study with German vowels [o], [&#x000F8;], and [e], which differ mainly in place of articulation. When discriminating [o]&#x02014;[&#x000F8;], larger electrophysiological responses occurred because of the mismatching features [DOR]&#x02014;[COR]. In the reverse direction, the effects were attenuated due to the underspecification of the coronal place of articulation. Similar results have been produced by Scharinger et al. (<xref ref-type="bibr" rid="B78">2012b</xref>) for tongue height oppositions using American English vowels for which the mid of tongue height is believed to be underspecified. They found larger effects if the mid vowel [&#x003B5;] had to be discriminated from low vowel [&#x000E6;] due to the mismatching features of [LOW] and [MID] compared to the reverse sequence, in which there is no feature mismatch due to underspecification. Similar evidence for this approach has been found not only for vowels (Lipski et al., <xref ref-type="bibr" rid="B40">2007</xref>; de Jonge and Boersma, <xref ref-type="bibr" rid="B12">2015</xref>) but also for consonants (Hestvik and Durvasula, <xref ref-type="bibr" rid="B27">2016</xref>; Schluter et al., <xref ref-type="bibr" rid="B80">2016</xref>, <xref ref-type="bibr" rid="B79">2017</xref>; Cummings et al., <xref ref-type="bibr" rid="B11">2017</xref>; H&#x000F8;jlund et al., <xref ref-type="bibr" rid="B29">2019</xref>; Hestvik et al., <xref ref-type="bibr" rid="B28">2020</xref>) and suprasegmental elements like lexical tones (Politzer-Ahles et al., <xref ref-type="bibr" rid="B61">2016</xref>). While most studies used isolated vowels or syllables, there is also evidence from complex stimuli like words (Friedrich et al., <xref ref-type="bibr" rid="B22">2008</xref>; Scharinger et al., <xref ref-type="bibr" rid="B78">2012b</xref>; Cornell et al., <xref ref-type="bibr" rid="B10">2013</xref>; Lawyer and Corina, <xref ref-type="bibr" rid="B39">2018</xref>).</p>
<p>The other model investigated in this article, the NRV framework, also predicts different discrimination performances as a function of presentation order. In contrast to the aforementioned model, NRV operationalizes phonetic properties of the speech signal which can be specified by acoustical or visual cues to explain directional asymmetries and predict different discrimination performances and proposes that &#x0201C;vowels with extreme articulatory-acoustic properties (peripheral in the vowel space;&#x0201D; Polka and Bohn, <xref ref-type="bibr" rid="B64">2011</xref>, p. 474) are so-called referent vowels and are easier to discriminate. Polka and Bohn (<xref ref-type="bibr" rid="B63">2003</xref>, <xref ref-type="bibr" rid="B64">2011</xref>) observed a universal perceptual bias favoring vowel discrimination from a more central to a more peripheral vowel in the vowel space in infants. They proposed that the vowels on the periphery of the vowel space (/i/, /a/, /u/, /y/) act as universal referent vowels in language development and vowel discrimination due to their more salient and extreme articulatory-acoustic properties. The vowel space periphery&#x02019;s perceptual advantage can be explained by the convergence of adjacent formants and therefore the stronger focalization of the referent vowels (Schwartz et al., <xref ref-type="bibr" rid="B84">1997</xref>, <xref ref-type="bibr" rid="B83">2005</xref>). Since this framework has been developed from the point of view of language acquisition and infant vowel discrimination, much work has been done on the investigation of the proposed perceptual bias in infants. There is evidence from an early cross-linguistic study with German- and English-learning infants that for English vowels /&#x003B5;/ and /&#x000E6;/, discrimination was easier for /&#x003B5;/&#x02014;/&#x000E6;/ than in the reverse direction, regardless of the language background of the infants (Polka and Bohn, <xref ref-type="bibr" rid="B62">1996</xref>). A similar bias with easier discrimination from a more central (less focal) to a more peripheral (more focal) vowel was shown in several studies (Bohn and Polka, <xref ref-type="bibr" rid="B7">2001</xref>; Polka and Bohn, <xref ref-type="bibr" rid="B64">2011</xref>; Pons et al., <xref ref-type="bibr" rid="B65">2012</xref>; Simon et al., <xref ref-type="bibr" rid="B89">2014</xref>). Additionally, there is some encouraging evidence that the perceptual bias preferring some sounds to others in discrimination in infants also could hold true in consonants (Nam and Polka, <xref ref-type="bibr" rid="B51">2016</xref>). Concerning adult vowel perception and discrimination, within the framework, it was initially proposed that the perceptual bias is shaped by language experience. Therefore, the asymmetry only occurs if subjects are discriminating non-native vowel contrasts, while in native vowel contrasts, the perceptual bias disappears, and asymmetry occurs (Polka and Bohn, <xref ref-type="bibr" rid="B64">2011</xref>). The assumption of experience-dependent asymmetries was found in some studies (Tyler et al., <xref ref-type="bibr" rid="B99">2014</xref>; Kriengwatana and Escudero, <xref ref-type="bibr" rid="B35">2017</xref>) while others also report universal biases in adults. In an AX discrimination test with Canadian-French and Canadian-English subjects using tokens of less focal English /u/ and more focal French /u/, Masapollo et al. (<xref ref-type="bibr" rid="B44">2017b</xref>) found that discrimination from less to more focalization produced better and faster results irrespective of language background. Therefore, the authors argued that there is a universal bias towards more focalized vowels in adults, too. These results have been replicated and extended in that the universal bias seems to have an impact not only on the auditory domain of speech processing but also on visual vowel discrimination (Masapollo et al., <xref ref-type="bibr" rid="B42">2017a</xref>, <xref ref-type="bibr" rid="B43">2018</xref>).</p>
<p>In recent research, mental representations of speech sounds have often been investigated with the help of electrophysiological methods, for example by using event-related potentials (ERPs). ERPs offer a means to investigate speech processing on a temporal axis with the accuracy of milliseconds. In the investigation of speech sound processing, one ERP component, the so-called Mismatch Negativity (MMN), has become prominent. The MMN can be defined as a specific electrophysiological detection change response of the brain when the repetitive presentation of one stimulus (standard) is interrupted occasionally and unpredictably by a different stimulus (N&#x000E4;&#x000E4;t&#x000E4;nen et al., <xref ref-type="bibr" rid="B48">2007</xref>). The MMN component has often been used for the investigation of (speech) sound processing since this component can be elicited even when participants are not attending to the stimulation. It, therefore, reflects preattentive and automatic speech processing, making it possible to differentiate the neural responses of stimuli without attention effects and other perceptual and cognitive processes (for a review on the component, see N&#x000E4;&#x000E4;t&#x000E4;nen et al., <xref ref-type="bibr" rid="B48">2007</xref>). This component usually peaks fronto-centrally between 100 and 250 ms after change onset and can be elicited by any discriminable change in the stimulation (N&#x000E4;&#x000E4;t&#x000E4;nen, <xref ref-type="bibr" rid="B50">2001</xref>), for example in pure tones (e.g., Sams et al., <xref ref-type="bibr" rid="B70">1985</xref>), with sensitivity for changes in frequency (for example Takegata and Morotomi, <xref ref-type="bibr" rid="B92">1999</xref>; Tervaniemi et al., <xref ref-type="bibr" rid="B96">2000</xref>), intensity and duration (e.g., Paavilainen et al., <xref ref-type="bibr" rid="B54">1991</xref>) but also in more complex stimuli like speech sounds (Dehaene-Lambertz, <xref ref-type="bibr" rid="B15">1997</xref>; Dehaene-Lambertz et al., <xref ref-type="bibr" rid="B14">2000</xref>). Furthermore, several studies have shown that the latency of the component is usually linked to the complexity of the stimuli, while the amplitude of the MMN is correlated with the magnitude of deviation. The greater the differences between the standard and the deviant stimulus are, the greater the MMN (e.g., Sams et al., <xref ref-type="bibr" rid="B70">1985</xref>; Savela et al., <xref ref-type="bibr" rid="B71">2003</xref>). Moreover, it has been shown that the MMN component is sensitive for language-specific phonemic processing of speech sounds (e.g., Dehaene-Lambertz, <xref ref-type="bibr" rid="B15">1997</xref>; N&#x000E4;&#x000E4;t&#x000E4;nen et al., <xref ref-type="bibr" rid="B49">1997</xref>), which led to the interpretation that mental representations are of phonemic or phonetic nature (as opposed to auditory ones) and language-specific.</p>
<p>In this article, we tested both competing models with German vowels to investigate which model can best explain directional asymmetries. Consequently, we tested the predictions of both models on a large stimulus set (five German long vowel contrasts) in a more natural listening situation by using real minimal pairs. The study was based on the following notions.</p>
<p>The use of real words in MMN investigations is associated with obstacles due to interference, such as lexical status, familiarity, or other confounding factors. In several studies, it has been shown that MMN responses for real word deviants are enhanced in comparison to pseudowords: it is believed that the enhancement of the lexical MMN is due to stronger memory trace activation for real meaningful words (Pulverm&#x000FC;ller et al., <xref ref-type="bibr" rid="B66">2001</xref>, <xref ref-type="bibr" rid="B68">2004</xref>; Shtyrov and Pulverm&#x000FC;ller, <xref ref-type="bibr" rid="B88">2002</xref>; Endrass et al., <xref ref-type="bibr" rid="B17">2004</xref>; Pettigrew et al., <xref ref-type="bibr" rid="B58">2004a</xref>; Shtyrov et al., <xref ref-type="bibr" rid="B87">2008</xref>). Another known influential factor on speech processing is the lexical frequency of the used real words. This influence can even be present when testing real words in a passive oddball paradigm and can lead to a stronger MMN response for words with higher lexical frequency in opposition to deviants with a lower or intermediate frequency of occurrence (Alexandrov et al., <xref ref-type="bibr" rid="B4">2011</xref>; Shtyrov et al., <xref ref-type="bibr" rid="B86">2011</xref>; Aleksandrov et al., <xref ref-type="bibr" rid="B3">2017</xref>). Furthermore, it has been shown that phonotactic probabilities (sequential order of phonemes in words) influence MMN results with higher probability, accompanied by enhanced MMN effects (Bonte et al., <xref ref-type="bibr" rid="B8">2005</xref>; Yasin, <xref ref-type="bibr" rid="B104">2007</xref>; Emmendorfer et al., <xref ref-type="bibr" rid="B16">2020</xref>). Concerning vowel perception, acoustic properties, like for example fundamental frequency, vowel duration or intensity (Aaltonen et al., <xref ref-type="bibr" rid="B1">1994</xref>; Kirmse et al., <xref ref-type="bibr" rid="B34">2007</xref>; Peter et al., <xref ref-type="bibr" rid="B57">2010</xref>; Partanen et al., <xref ref-type="bibr" rid="B56">2011</xref>), have an impact on neural effects. While some of the mentioned influential factors can be controlled for when developing stimulus materials, others are not avoidable. For instance, various acoustic differences in vowels stem from collinearities between vowel identity and acoustic consequences. Changes in vowel identity simultaneously lead to changes in the spectral frequency structure (mainly F1 and F2) of the stimuli. Moreover, vowel features used in theoretical frameworks are based largely on articulatory-acoustic properties&#x02014;mainly formants&#x02014;and therefore, it could also be possible that there is more of an acoustical influence, especially in MMN effects, than proposed by operationalizing more abstract and theoretical derived features. For instance, the feature opposition of [HIGH] and [LOW] is the more abstract representation of the articulatory and acoustic properties of those vowels concerning the first formant: high vowels have a low F1, while low vowels have a high F1 (Lahiri and Reetz, <xref ref-type="bibr" rid="B37">2010</xref>). Also, the abstract description of vowels referring to focality which were used in the NRV framework is based on articulatory-acoustic properties, since focalization stems from the convergence of adjacent formants (Schwartz et al., <xref ref-type="bibr" rid="B84">1997</xref>). While the common contributing articulatory-based factors of vowel perception have often been investigated, the influence of perceptual and psychoacoustic parameters (e.g., perceptual loudness) on vowel perception has hardly been studied. Thus, we wanted to additionally investigate which were the influential factors on vowel discrimination, including not only theoretical and acoustical factors but also perceptual factors beyond the well-known.</p>
<p>Hence, the following research questions shall be investigated: (1) which model accommodates directional asymmetries in the processing of natural and unmanipulated German long vowels in the best way; and (2) which factors influence vowel discrimination in natural German minimal pairs? The first question has been addressed on an electrophysiological level through measurement of MMN (Experiment 1) and on a behavioral level in means of reaction times (RT; Experiment 2). The second aim of identifying influential factors on vowel discrimination, pursued <italic>via</italic> multiple regressions on both datasets, should shed more light on factors that co-determine MMN effects.</p>
</sec>
<sec id="s2">
<title>Experiment 1: Mmn Study</title>
<p>To test both models, we first conducted an MMN study with a large stimulus set, testing five German long vowel contrasts embedded in natural minimal pairs, which almost mapped the entire German (long) vowel space. The vowels chosen for investigation were among the most frequent long vowels in the German language (Aichert et al., <xref ref-type="bibr" rid="B2">2005</xref>).</p>
<sec id="s2-1">
<title>Participants</title>
<p>Nineteen participants (nine females, mean age 24.7, SD 3.4), graduate and undergraduate students of the Philipps University of Marburg, participated in two sessions for monetary compensation. They were all right-handed and reported no hearing or neurological impairments. All participants were monolingual German native and German Standard speakers without being able to speak any German dialect actively. They were all born and socialized in Hesse, Germany, with Standard German. The information about the participants&#x02019; dialect and Standard German competence was retrieved by questionnaire. Informed written consent was obtained from each participant before the experiment. One subject had to be excluded because the participant missed the second session. Another subject had to be excluded due to excessive contamination with artifacts in the EEG data (movement artifacts). In total, we assessed and analyzed the complete data of 17 participants.</p>
</sec>
<sec sec-type="materials" id="s2-2">
<title>Materials</title>
<p>To test the hypotheses of the aforementioned models, we chose the five German long vowel contrasts /i:/&#x02014;/e:/, /e:/&#x02014;/a:/, /y:/&#x02014;/u:/, /i:/&#x02014;/u:/, and /i:/&#x02014;/a:/. They differed concerning the place of articulation, vowel height as well as rounding. To ensure more phonological processing, we embedded these vowels in German monosyllabic minimal pairs. We tried to keep the phonetic context between pairs as similar as possible. We also controlled for the frequency of occurrence with SUBTlex (Brysbaert et al., <xref ref-type="bibr" rid="B9">2011</xref>), as seen in <xref ref-type="table" rid="T1">Table 1</xref>.</p>
<table-wrap id="T1" position="float">
<label>Table 1</label>
<caption><p>Phonetic and lexical parameters of the vowels.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="center">Vowel contrast</th>
<th align="center">Words F0 (SD)</th>
<th align="center">Mean F1 (SD)</th>
<th align="center">Mean F2 (SD)</th>
<th align="center">Mean F3 (SD)</th>
<th align="center">Mean intensity (SD)</th>
<th align="center">Mean duration (ms)</th>
<th align="center">Mean (log-values)</th>
<th align="center">Word frequency</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">/a:/&#x02014;/i:/</td>
<td align="center"><italic>Zahl</italic> (&#x0201C;number&#x0201D;)</td>
<td align="center">169 (24)</td>
<td align="center">875 (19)</td>
<td align="center">1,488 (59)</td>
<td align="center">3,046 (77)</td>
<td align="center">72.85 (0.27)</td>
<td align="center">292</td>
<td align="center">2.861</td>
</tr>
<tr>
<td/>
<td align="center"><italic>Ziel</italic> (&#x0201C;target&#x0201D;)</td>
<td align="center">203 (5)</td>
<td align="center">244 (28)</td>
<td align="center">2,445 (61)</td>
<td align="center">3,446 (74)</td>
<td align="center">73.79 (0.92)</td>
<td align="center">159</td>
<td align="center">3.358</td>
</tr>
<tr>
<td align="left">/e:/&#x02014;/i:/</td>
<td align="center"><italic>Steg</italic> (&#x0201C;bridge&#x0201D;)</td>
<td align="center">179 (10)</td>
<td align="center">359 (6)</td>
<td align="center">2,483 (58)</td>
<td align="center">3,171 (113)</td>
<td align="center">73,96 (0.67)</td>
<td align="center">270</td>
<td align="center">1.255</td>
</tr>
<tr>
<td/>
<td align="center">Stieg (&#x0201C;climbed&#x0201D;)</td>
<td align="center">190 (3)</td>
<td align="center">286 (11)</td>
<td align="center">2,479 (51)</td>
<td align="center">3,528 (21)</td>
<td align="center">74.92 (0.31)</td>
<td align="center">228</td>
<td align="center">2.352</td>
</tr>
<tr>
<td align="left">/a:/&#x02014;/e:/</td>
<td align="center"><italic>Mahl</italic> (&#x0201C;meal&#x0201D;)</td>
<td align="center">174 (5)</td>
<td align="center">913 (10)</td>
<td align="center">1,484 (47)</td>
<td align="center">2,966 (44)</td>
<td align="center">71.68 (0.22)</td>
<td align="center">225</td>
<td align="center">1.672</td>
</tr>
<tr>
<td/>
<td align="center"><italic>Mehl</italic> (&#x0201C;flour&#x0201D;)</td>
<td align="center">191 (3)</td>
<td align="center">341 (8)</td>
<td align="center">2,566 (39)</td>
<td align="center">3,377 (209)</td>
<td align="center">71.74 (0.26)</td>
<td align="center">195</td>
<td align="center">1.857</td>
</tr>
<tr>
<td align="left">/u:/&#x02014;/i:/</td>
<td align="center"><italic>Stuhl</italic> (&#x0201C;chair&#x0201D;)</td>
<td align="center">200 (8)</td>
<td align="center">294 (37)</td>
<td align="center">1,974 (212)</td>
<td align="center">2,647 (99)</td>
<td align="center">73.88 (0.44)</td>
<td align="center">186</td>
<td align="center">2.892</td>
</tr>
<tr>
<td/>
<td align="center"><italic>Stiel</italic> (&#x0201C;handle&#x0201D;)</td>
<td align="center">199 (6)</td>
<td align="center">275 (22)</td>
<td align="center">2,475 (75)</td>
<td align="center">3,579 (59)</td>
<td align="center">73.67 (0.44)</td>
<td align="center">176</td>
<td align="center">1.756</td>
</tr>
<tr>
<td align="left">/u:/&#x02014;/y:/</td>
<td align="center"><italic>Sud</italic> (&#x0201C;brew&#x0201D;)</td>
<td align="center">194 (2)</td>
<td align="center">296 (40)</td>
<td align="center">1,145 (140)</td>
<td align="center">2,628 (93)</td>
<td align="center">74.75 (0.20)</td>
<td align="center">194</td>
<td align="center">0.845</td>
</tr>
<tr>
<td/>
<td align="center"><italic>S&#x000FC;d</italic> (&#x0201C;south&#x0201D;)</td>
<td align="center">180 (5)</td>
<td align="center">292 (14)</td>
<td align="center">2,085 (120)</td>
<td align="center">2,633 (146)</td>
<td align="center">74.43 (0.43)</td>
<td align="center">223</td>
<td align="center">1.771</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>Mean F0, F1, F2, and F3 values are given in Hertz for vowels per word category. Mean Intensity for the vowels within the words are given in dB. Mean duration measures referring to the vowels within the words. The frequency of occurrence (as log-values) for each word is given in the last column</italic>.</p>
</table-wrap-foot>
</table-wrap>
<p>Twenty natural exemplars of each word were recorded in a sound shielded booth by a female German Standard speaker who was phonetically trained. All tokens were spoken with neutral pronunciation. All sounds have been analyzed for F0, F1, F2, F3, as well as vowel duration and were all scaled to an intensity level of 70 dB within Praat (Boersma and Weenink, <xref ref-type="bibr" rid="B6">2016</xref>). The five best tokens per word have been chosen as experimental stimuli. Phonetic parameters of the word categories are displayed in <xref ref-type="table" rid="T1">Table 1</xref>. Note that we reported here only mean values per word category (since MMN and RT data are also averaged measures), but a more detailed description of the acoustic parameters can be found in <xref ref-type="supplementary-material" rid="SM1">Supplementary Table 1</xref>. There can be seen that our stimuli had some variance regarding, for example, vowel duration. Since, we wanted to test natural spoken words, no manipulation was applied. All vowels should be perceived as long vowels despite the length differences since the phonological category is additionally supported by the lexical context. Therefore, the focus in processing lies on categorical differences regarding vowel height and place of articulation. All experimental stimuli were found to sound natural by two different persons. All tokens were also assessed as being distinct for their category (see <xref ref-type="fig" rid="F1">Figure 1</xref>). We compared the formant values (F1, F2) to the ones of Sendlmeier and Seebode (<xref ref-type="bibr" rid="B85">2006</xref>) to ensure that they will be perceived as Standard German. We chose to introduce inter-token variation to obtain a more natural listening situation and to ensure a more phonological approach since participants are forced to map the incoming variable acoustic signals onto a unified and more abstract representation to cope with inter-token variability (Phillips et al., <xref ref-type="bibr" rid="B60">2000</xref>; Eulitz and Lahiri, <xref ref-type="bibr" rid="B20">2004</xref>; Jacobsen et al., <xref ref-type="bibr" rid="B31">2004</xref>). This is an important design feature since it mediates the likely collinearity between formant frequencies and acoustic- or articulatory-phonetic features.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>Acoustic characteristic of the stimuli. Mean values of the first (F1) and the second (F2) formant are given per word category in Hertz.</p></caption>
<graphic xlink:href="fnhum-15-612345-g0001.tif"/>
</fig>
</sec>
<sec id="s2-3">
<title>Task and Procedure</title>
<p>The stimuli were embedded in a passive oddball design. In this paradigm, the participants were presented with a series of repetitive stimuli (standards) that were interspersed occasionally by a deviant varying only in vowel quality while they were watching a silent movie. The frequently presented standards were assumed to activate the memory trace and therefore the representation in the mental lexicon, whereas the infrequently presented deviants provided information about and are processed closer to the surface structure. Each vowel contrast was tested bidirectionally. Because, we investigated five contrasts in both directions, all subjects were tested in two sessions (with testing times per session approximating 2 h) within 15&#x02013;20 days. Thus, each word served as standard and as deviant in different blocks and sessions.</p>
<p>Each contrast direction was presented in two blocks containing 425 standards and 75 deviants each. In total, we presented 850 standards and 150 deviants per contrast direction. Thus, we presented 2,000 stimuli for each vowel contrast. Within the blocks, stimuli were randomized, and the interval between two deviants randomly consisted of 4&#x02013;11 standards. Blocks were randomized for both sessions. Blocks of the same condition never succeeded each other. The fixed ISI was 1,000 ms, while the stimuli varied in duration. Therefore, we still obtained a jittered presentation to suppress rhythmic processing and habituation to synchronously presented stimuli.</p>
<p>Subjects were seated comfortably in a sound-insulated and electromagnetically shielded chamber in front of a screen. Sounds were presented binaurally at a comfortable listening level <italic>via</italic> two loudspeakers on the left and the right of the screen, using the open-source software OpenSesame (Math&#x000F4;t et al., <xref ref-type="bibr" rid="B45">2012</xref>). The listening level was set before the experiment and was kept equal across all subjects (based on the intensity level of 70 dB as manipulated in PRAAT).</p>
</sec>
<sec id="s2-4">
<title>Hypotheses</title>
<p>Since this article aims at comparing the two aforementioned models, hypotheses have been made on the assumptions of vowel discrimination following FUL as well as NRV. For NRV we proposed the hypothesis based on the universal assumptions of the framework that vowels /i:/, /y:/, /u:/, and /a:/ are reference vowels (Polka and Bohn, <xref ref-type="bibr" rid="B64">2011</xref>). The basic assumptions of both models regarding feature specifications and position in the vowel space for each investigated contrast are displayed in <xref ref-type="table" rid="T2">Table 2</xref>.</p>
<table-wrap id="T2" position="float">
<label>Table 2</label>
<caption><p>Assumptions for feature specifications (FUL) and location in the vowel space (NRV) for each vowel contrast and hypotheses for mismatch negativity (MMN) effects according to both models.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="center">Vowel contrast</th>
<th align="center">Presentation order</th>
<th align="center">Features (FUL)</th>
<th align="center">Mapping result (FUL)</th>
<th align="center">Expectations MMN (FUL)</th>
<th align="center">Classification (NRV)</th>
<th align="center">Expectations MMN (NRV)</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">/a:/&#x02014;/i:/</td>
<td align="center"><italic>Zahl&#x02014;Ziel</italic></td>
<td align="center">[DOR]&#x02014;[COR] [LOW]&#x02014;[HIGH]</td>
<td align="center">Mismatch Mismatch</td>
<td align="center">Stronger effect</td>
<td align="center">Both peripheral</td>
<td align="center">Symmetrical effect</td>
</tr>
<tr>
<td/>
<td align="center"><italic>Ziel&#x02014;Zahl</italic></td>
<td align="center">[&#x02013;]&#x02014;[DOR] [HIGH]&#x02014;[LOW]</td>
<td align="center">No-mismatch Mismatch</td>
<td align="center">Weaker effect</td>
<td/>
<td/>
</tr>
<tr>
<td align="left">/e:/&#x02014;/i:/</td>
<td align="center"><italic>Steg&#x02014;Stieg</italic></td>
<td align="center">[&#x02013;]&#x02014;[HIGH]</td>
<td align="center">No-mismatch</td>
<td align="center">Weaker effect</td>
<td align="center">Central&#x02014;peripheral</td>
<td align="center">Stronger effect</td>
</tr>
<tr>
<td/>
<td align="center"><italic>Stieg&#x02014;Steg</italic></td>
<td align="center">[HIGH]&#x02014;[MID]</td>
<td align="center">Mismatch</td>
<td align="center">Stronger effect</td>
<td align="center">Peripheral&#x02014;central</td>
<td align="center">Weaker effect</td>
</tr>
<tr>
<td align="left">/a:/&#x02014;/e:/</td>
<td align="center"><italic>Mahl&#x02014;Mehl</italic></td>
<td align="center">[DOR]&#x02014;[COR] [LOW]&#x02014;[MID]</td>
<td align="center">Mismatch Mismatch</td>
<td align="center">Stronger effect</td>
<td align="center">Peripheral&#x02014;central</td>
<td align="center">Weaker effect</td>
</tr>
<tr>
<td/>
<td align="center"><italic>Mehl&#x02014;Mahl</italic></td>
<td align="center">[&#x02013;]&#x02014;[DOR] [&#x02013;]&#x02014;[LOW]</td>
<td align="center">No-mismatch No-mismatch</td>
<td align="center">Weaker effect</td>
<td align="center">Central&#x02014;peripheral</td>
<td align="center">Stronger effect</td>
</tr>
<tr>
<td align="left">/u:/&#x02014;/i:/</td>
<td align="center"><italic>Stuhl&#x02014;Stiel</italic></td>
<td align="center">[DOR]&#x02014;[COR]</td>
<td align="center">Mismatch</td>
<td align="center">Stronger effect</td>
<td align="center">Both peripheral</td>
<td align="center">Symmetrical effect</td>
</tr>
<tr>
<td/>
<td align="center"><italic>Stiel&#x02014;Stuhl</italic></td>
<td align="center">[&#x02013;]&#x02014;[DOR]</td>
<td align="center">No-mismatch</td>
<td align="center">Weaker effect</td>
<td/>
<td/>
</tr>
<tr>
<td align="left">/u:/&#x02014;/y:/</td>
<td align="center"><italic>Sud&#x02014;S&#x000FC;d</italic></td>
<td align="center">[DOR]&#x02014;[COR]</td>
<td align="center">Mismatch</td>
<td align="center">Stronger effect</td>
<td align="center">Both peripheral</td>
<td align="center">Symmetrical effect</td>
</tr>
<tr>
<td/>
<td align="center"><italic>S&#x000FC;d&#x02014;Sud</italic></td>
<td align="center">[&#x02013;]&#x02014;[DOR]</td>
<td align="center">No-mismatch</td>
<td align="center">Weaker effect</td>
<td/>
<td/>
</tr>
</tbody>
</table>
</table-wrap>
<p>In accordance with the models, we predict the following MMN effects (see also <xref ref-type="table" rid="T2">Table 2</xref>): within FUL, the effects should be stronger for mismatching presentation orders /a:/&#x02014;/i:/, /u:/&#x02014;/y:/ and /u:/&#x02014;/i:/ (because of the mismatching features [DOR] and [COR]), /i:/&#x02014;/e:/ (due to mismatch of [HIGH] and [MID]) as well as /a:/&#x02014;/e:/ (mismatch of [DOR]&#x02014;[COR] and [HIGH]&#x02014;[MID]). If NRV holds true, MMN effects should be stronger when <italic>Stieg</italic> and <italic>Mahl</italic> are deviants since they are referent vowels in this models. In the other three contrasts, a symmetry should occur since both vowels are peripheral and act as referents within the framework.</p>
</sec>
<sec id="s2-5">
<title>EEG Recording and Analysis</title>
<p>EEG was recorded with 28 Ag/AgCl passive electrodes connected to a BrainAmp amplifier (Brain Products GmbH). Electrodes were arranged on an EasyCap in 10-20 positions. AFz served as the Ground electrode, and the online reference was placed on the nose tip. Four additional electrodes measured the electrooculogram (EOG) for the identification of artifacts caused by eye movements (e.g., blinks). Two electrodes were placed left and right of the eye canthi to measure lateral eye movements. Two electrodes above and under the right eye measured vertical eye movements. For all electrodes, impedances were kept below 5 k&#x003A9; and the sampling rate was 500 Hz.</p>
<p>EEG analysis was done with the MATLAB toolbox fieldtrip. Raw data were filtered with 0.16 and 30 Hz high- and low-pass filters. Data were re-referenced offline to linked mastoids. After segmentation, EEG data were automatically corrected for muscle artifacts. Eye movements were automatically corrected through the correlation of EOG channels and ICA components. The calculation of the MMN component was based on the onset of the vowel, i.e., epochs beginnings were aligned with vowel beginnings. Thereby, consonant onset clusters in the stimuli should play no role in the MMN effects. Additionally, ERP data were baseline corrected using the 100-ms prestimulus epoch.</p>
<p>For averaging the first ten standards of a block and the first standard after a deviant were excluded from data analysis. To maintain ERP results without the influence of pure acoustic influences, we calculated and plotted the MMN as identity MMN (iMMN). Here, the standard and the deviant of the same word are compared to each other (Pulverm&#x000FC;ller et al., <xref ref-type="bibr" rid="B67">2006</xref>).</p>
</sec>
<sec sec-type="results" id="s2-6">
<title>Results</title>
<p>The results of the iMMN study are plotted in <xref ref-type="fig" rid="F2">Figure 2</xref>.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Identity mismatch negativity (MMN) effects per condition. MMN waveforms for all word pairs, in both presentation orders, are shown.</p></caption>
<graphic xlink:href="fnhum-15-612345-g0002.tif"/>
</fig>
<p>In a first step, we were interested in significant standard-deviant differences in the auditory evoked potentials. To this end, we employed a conservative measure of amplitude contrasts without prior assumptions of regions of interest and followed a multilevel statistical approach (e.g., Henry and Obleser, <xref ref-type="bibr" rid="B26">2012</xref>; Strau&#x000DF; et al., <xref ref-type="bibr" rid="B90">2014</xref>). At the first level, we calculated independent-samples <italic>t-</italic>tests between the single-trial amplitude values of standards and deviants. Uncorrected by-participant <italic>t</italic>-values were obtained for all time-amplitude bins of all electrodes. At the second level, <italic>t</italic>-values were tested against 0 with dependent-sample <italic>t-</italic>tests. Taking into consideration the problem of multiple comparisons, a Monte-Carlo nonparametric permutation method with 1,000 randomizations, as implemented in fieldtrip (Oostenveld et al., <xref ref-type="bibr" rid="B53">2011</xref>), estimated type I-error controlled cluster significance probabilities (at <italic>p</italic> &#x0003C; 0.05). In an electrode &#x000D7; time cluster (with Fz, Cz, CPz, between 130 and 200 ms post vowel onset), deviants elicited a significantly more negative response than standards (see <xref ref-type="fig" rid="F3">Figure 3</xref>).</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Clusterstatistics. In an electrode &#x000D7; time cluster, deviants elicited more negative responses than standards in the time window between 130 and 200 ms post vowel onset.</p></caption>
<graphic xlink:href="fnhum-15-612345-g0003.tif"/>
</fig>
<p>To analyze the EEG data for directional asymmetries, we calculated the iMMN as difference waves (deviant minus standard of the same words) in this aforementioned time window. Then, we calculated repeated-measures and Bonferroni corrected ANOVAs for each contrast with factors <italic>word</italic> (e.g., <italic>Ziel</italic> vs. <italic>Zahl</italic>) and <italic>electrode</italic> (Fz, Cz, CPz). Electrodes were chosen by cluster statistics.</p>
<p>In the /i:/&#x02014;/a:/ contrast, we found both main effects for word (<italic>F</italic><sub>(1,16)</sub> = 7.286, <italic>p</italic> = 0.016) with a larger MMN for <italic>Zahl</italic> (<italic>M</italic> = &#x02212;1,721, SEM = 0.36; <italic>Ziel</italic>: <italic>M</italic> = &#x02212;0.586, SEM = 0.253) and for electrode (<italic>F</italic><sub>(2,32)</sub> = 14.634, <italic>p</italic> &#x0003C; 0.001) with strongest effect at Cz (<italic>F</italic><sub>(1,16)</sub> = 5.890, <italic>p</italic> &#x0003C; 0.05). In the vowel contrast /e:/&#x02014;/a:/, there was not only a highly significant main effect for electrode (<italic>F</italic><sub>(2,32)</sub> = 12.307, <italic>p</italic> &#x0003C; 0.001) but also an interaction word &#x000D7; electrode (<italic>F</italic><sub>(2,32)</sub> = 4.942, <italic>p</italic> = 0.013). <italic>Post hoc</italic> analysis of the interaction showed a significant effect on Cz (<italic>F</italic><sub>(1,16)</sub> = 5.039, <italic>p</italic> &#x0003C; 0.05). Hence, we found asymmetries in visual inspection as well as in statistical analysis in both contrasts. The comparison of /u:/&#x02014;/y:/ only revealed a main effect of factor electrode (<italic>F</italic><sub>(2,32)</sub> = 12.349, <italic>p</italic> &#x0003C; 0.001) with a marginal effect on CPz (<italic>F</italic><sub>(1,16)</sub> = 4.265, <italic>p</italic> = 0.055). Therefore, comparing <italic>S&#x000FC;d</italic> and <italic>Sud</italic>, we found an asymmetry in the visual inspection, which did not hold out statistical analysis. Hence, statistically we found a symmetrical effect.</p>
<p>The vowel contrast /i:/&#x02014;/e:/ shows a symmetrical pattern in visual inspection and statistics with both main effects insignificant (word: <italic>F</italic><sub>(1,16)</sub> = 1.687, <italic>p</italic> = 0.212, electrode: <italic>F</italic><sub>(2,32)</sub> = 2.367, <italic>p</italic> = 0.110). The same is also true for the comparison of /i:/&#x02014;/u:/ (word: <italic>F</italic><sub>(1,16)</sub> = 0.294, <italic>p</italic> = 0.595, electrode: <italic>F</italic><sub>(2,32)</sub> = 0.725, <italic>p</italic> = 0.492).</p>
</sec>
<sec sec-type="discussion" id="s2-7">
<title>Discussion</title>
<p>In summary, we found no clear evidence for neural asymmetries due to underspecification (FUL) but evidence for vowel discrimination based on phonetic salience of referent vowels (NRV). Furthermore, there were asymmetric as well as symmetric patterns in the MMN.</p>
<p>The asymmetric pattern of the comparison between /e:/&#x02014;/a:/ was in line with the hypothesis of NRV. Here, /a:/ is a referent vowel in addition to being more peripheral, and discrimination of /e:/&#x02014;/a:/ is, therefore, easier and comes with a stronger MMN effect than vice versa. Additionally, the symmetric effect in the contrast /y:/&#x02014;/u:/ can also be explained with this model since both vowels are referents within this framework. The same holds good for the comparison of MMN effects between presentation orders of /i:/&#x02014;/u:/. But in the latter contrast, there could be also a phonological explanation within the underspecification approach. The phonological variation in morphological processes can lead to different specifications of segments within words and therefore to effects that are at first sight not compatible within the FUL paradigm (Lawyer and Corina, <xref ref-type="bibr" rid="B39">2018</xref>). The same is true for German umlauting back vowels. In our case, when deriving the plural of the German word <italic>Stuhl</italic>, the stem vowel is umlauting and fronting (<italic>St&#x000FC;hle</italic>). It can be assumed that umlaut is only possible if the stem vowel /u:/ is not specified for the place of articulation features and is therefore underspecified for backness (Scharinger, <xref ref-type="bibr" rid="B72">2009</xref>; Scharinger et al., <xref ref-type="bibr" rid="B76">2010</xref>). If the stem vowel of <italic>Stuhl</italic> is underspecified for the place of articulation information, asymmetry has to occur when it is compared to /i:/, which is also underspecified.</p>
<p>Contrary to this, the results of the remaining two contrasts are somewhat challenging since none of the previous operationalized models can explain the effects given in the data. Comparing the presentation orders in the contrast /i:/&#x02014;/e:/, an asymmetric MMN pattern occurred. This is challenging the predictions of FUL as well as NRV since both models predict an asymmetry. According to the underspecification approach, MMN effects for <italic>Steg</italic> should be stronger (underspecification of mid vowel height), while NRV predicted that neural effects for <italic>Stieg</italic> should be stronger (/i:/ should act as focal referent here). An explanation for the symmetric effect could lie in the close phonetic distance of the vowels involved. There has been evidence from previous MMN studies that effects diminished or failed due to only small acoustic deviances in speech stimuli (Pettigrew et al., <xref ref-type="bibr" rid="B58">2004a</xref>, <xref ref-type="bibr" rid="B59">b</xref>).</p>
<p>The most challenging results were obtained in the vowel contrast /i:/&#x02014;/a:/. Although there is an asymmetric effect, both models failed to predict the direction of the found asymmetry: FUL predicted stronger effects for <italic>Ziel</italic> as coronal deviant due to mismatching place of articulation (PoA) information. In comparison with <italic>Zahl</italic> as standard, which is classified as a dorsal vowel (Scharinger, <xref ref-type="bibr" rid="B72">2009</xref>), the extracted feature [COR] from the acoustic signal of /i:/ should evoke a mismatching stronger MMN. Also, the mismatching height features of those two vowels cannot have evoked the asymmetry. Since both height features ([HIGH] and [LOW]) involved are specified in the underlying representation, a mismatch occurs regardless of the presentation order. Since a mismatch of those features occurs in both presentation orders, they should not evoke an asymmetric effect. Additionally, the NRV model cannot explain the found asymmetric pattern either. According to NRV, asymmetry should have occurred since both vowels act as focal referents within this framework. The explanation for these results is still unclear. We argue that since the more abstract feature representations are based on acoustic properties (mainly formants), the effects could be more driven by changes in the acoustics than in feature representations. Because this is the contrast with the largest difference in terms of F1 or degree of openness, spectral characteristics (e.g., changes in F1) of the vowels could have been more involved in eliciting the surprising effects on an automatic and preattentive level. Additionally, changes in vowel quality do not only lead to changes of formants but also result in changes in other perceptual and psychoacoustic parameters. There is evidence that, for example, the perceived loudness of speech stimuli varies for vowel quality. That is, lower front vowels are perceived louder despite equal intensity (Glave and Rietveld, <xref ref-type="bibr" rid="B23">1975</xref>, <xref ref-type="bibr" rid="B24">1979</xref>) and vocal effort (Eriksson and Traunm&#x000FC;ller, <xref ref-type="bibr" rid="B18">1999</xref>, <xref ref-type="bibr" rid="B19">2002</xref>). Thus, we hypothesize that psychoacoustic and perceptual parameters such as perceived loudness could have played a crucial role. This possibility is explored in greater detail using multiple regression analyses in &#x0201C;Explorative Analysis for Additional Influential Factors in MMN and log RT Data&#x0201D; section.</p>
</sec>
</sec>
<sec id="s3">
<title>Experiment 2: Reaction Times</title>
<p>Since our MMN results present some evidence that effects were not only driven by phonemic factors but also by acoustic differences, we decided to conduct a RT study in an attended listening task. It has been shown that the MMN evoked by unattended processing is sensitive to a great variety of different dimensions between standard and deviant. Here, preattentive processing has been proven to be sensitive also for low-level information like variations in duration and intensity (N&#x000E4;&#x000E4;t&#x000E4;nen et al., <xref ref-type="bibr" rid="B47">1989</xref>; Paavilainen et al., <xref ref-type="bibr" rid="B54">1991</xref>; Schr&#x000F6;ger, <xref ref-type="bibr" rid="B82">1996</xref>; Jacobsen et al., <xref ref-type="bibr" rid="B32">2003</xref>) or acoustic distance of stimuli (Savela et al., <xref ref-type="bibr" rid="B71">2003</xref>; Deguchi et al., <xref ref-type="bibr" rid="B13">2010</xref>). Since this component is highly sensitive to small low-level information differences (i.e., changes in frequency), higher-order information (for example phonemic identity) may be ignored or overridden in preattentive processing, for example, by acoustic proximity (Pettigrew et al., <xref ref-type="bibr" rid="B58">2004a</xref>). Therefore, RT in an active discrimination task might reflect more cognitive, decision-based processing in which higher and more abstract effects like phonemic discrimination might surface better and more clearly. For this study, we thus propose the same hypotheses regarding potential asymmetries for both models as in Experiment 1.</p>
<sec id="s3-1">
<title>Participants and Materials</title>
<p>Twenty-six participants (17 females, mean age 24.43, SD 4.23) were recruited, all of whom were graduate or undergraduate students at Johannes Gutenberg University Mainz. They received monetary compensation for their efforts. All participants were right-handed monolingual German speakers with no active dialect competence and were socialized with Standard German. No participant reported neurological, psychological, or hearing impairments. Written informed consent was obtained from each participant before the experiment.</p>
<p>The stimuli used in Experiment 2 were the same as in Experiment 1. In contrast to the prior experiment, we had to reduce the number of tested vowel contrasts in order to shorten the session length (approximately testing 45 min). Therefore, we tested only the vowel contrasts /i:/&#x02014;/a:/, /i:/&#x02014;/e:/, and /i:/&#x02014;/u:/. Contrasts were chosen as followed: /i:/&#x02014;/e:/ and /i:/&#x02014;/u:/ obtained in the MMN investigation symmetrical patterns and /i:/&#x02014;/a:/ evoked an asymmetrical pattern. The symmetrical pattern of /i:/&#x02014;/u:/ could be explainable with NRV and will therefore serve as control contrast for the remaining two vowel oppositions. Here, the iMMN results were not explainable by either of the models.</p>
</sec>
<sec id="s3-2">
<title>Task and Procedure</title>
<p>Stimuli were presented in an active oddball setup, in which participants had to press a button as soon as they perceived the deviant. They were told to perform a categorical, phonemic decision (and therefore ignoring the inter-token variability; Johnson, <xref ref-type="bibr" rid="B33">2015</xref>). During the experiment, subjects were seated comfortably in front of a screen in a sound shielded chamber. Sounds were presented with the Presentation software (version 16.4)<xref ref-type="fn" rid="fn0001"><sup>1</sup></xref> at a comfortable volume <italic>via</italic> two loudspeakers to the left and right of the screen. The volume was set prior to the experiment and was kept equal across all subjects. All written instructions were presented on screen. This way, participants were also informed about the beginning and end of the experiment as well as pauses. Each contrast direction contained 180 standards and 20 deviants divided into two blocks. In total, 12 blocks were presented. Stimuli within blocks were randomized with 4&#x02013;11 standards between two deviants. Blocks of the same condition never followed each other.</p>
</sec>
<sec id="s3-3">
<title>Analysis and Results</title>
<p>The reaction time analysis was based on correct responses only (98% of data points included). RT were corrected for the onset cluster of each stimulus. Thus, measurement of RT began on the vowel onset. RT faster than 100 ms and slower than 1,000 ms were excluded. The remaining data were log-transformed to obtain an approximately normal distribution (Ratcliff, <xref ref-type="bibr" rid="B69">1993</xref>; Whelan, <xref ref-type="bibr" rid="B100">2008</xref>). Outliers (&#x000B1;2.5 SD) were removed before statistical analysis.</p>
<p>A repeated measures ANOVA with the factor <italic>word</italic> (e.g., <italic>Ziel</italic> vs. <italic>Zahl</italic>), controlled for multiple testing by applying Bonferroni correction, was calculated to reveal possible behavioral asymmetries. Here, we found a highly significant main effect (<italic>F</italic><sub>(5,2,320)</sub> = 107.811, <italic>p</italic> &#x0003C; 0.001). <italic>Post hoc</italic> analysis revealed asymmetric patterns in two of the tested vowel contrasts. /i:/&#x02014;/e:/ was significantly faster than vice versa (<italic>F</italic><sub>(1,464)</sub> = 22.234, <italic>p</italic> &#x0003C; 0.001). The same was true for /i:/&#x02014;/u:/ (<italic>F</italic><sub>(1,464)</sub> = 13.550, <italic>p</italic> &#x0003C; 0.001). However, in the vowel contrast /i:/&#x02014;/a:/ a symmetrical pattern of RT occurred. Here, the <italic>post hoc</italic> analysis shows no difference between presentation directions (<italic>F</italic><sub>(1,464)</sub> = 0.793, <italic>p</italic> = 0.374; see <xref ref-type="fig" rid="F4">Figure 4</xref>).</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Reaction time results per condition. Reaction time results are given as log values per presentation direction of words with whiskers indicating the variance of the data and small dots representing outliers (but not extreme values) which were beneath the &#x000B1;2 SD cut-off.</p></caption>
<graphic xlink:href="fnhum-15-612345-g0004.tif"/>
</fig>
</sec>
<sec sec-type="discussion" id="s3-4">
<title>Discussion</title>
<p>The behavioral study aimed to investigate the basis for some of the electrophysiological found effects in Experiment 1. There, both models could not provide a comprehensive explanation for our results, meaning that both models failed to explain all found effects. The vowel contrasts /i:/&#x02014;/a:/ and /i:/&#x02014;/e:/ were particularly challenging. Therefore, we conducted a reaction time experiment in an active oddball paradigm to investigate the previously found neural patterns in more detail. Overall, the RT indicate that in an active discrimination paradigm, German natural word stimuli were discriminated phonemically, based on higher-order abstract phonological features.</p>
<p>The RT for the vowel contrast /i:/&#x02014;/e:/ and the observed asymmetrical pattern match with the predictions of FUL. The faster RT obtained when /e:/ was deviant seem to be due to the underspecification of [COR]. In this contrast, abstract representations may help to discriminate concerning the close phonetic distance of the vowels. There is evidence from an fMRI study indicating that participants had to rely more on abstract feature representations while discriminating acoustically very close vowels (Scharinger et al., <xref ref-type="bibr" rid="B74">2016</xref>).</p>
<p>In contrast, the directional asymmetry of the vowel contrasts /i:/&#x02014;/u:/ is, on first sight, more challenging for the underspecification approach since, following the theory, RT should be faster when subjects are presented with a fully specified vowel (e.g., /u:/) followed by an underspecified vowel (e.g., /i:/). But concerning the present study, we found the opposite effect for the obtained RT. The hypothesis of NRV for this contrast seems equally unsuitable: it states that asymmetric effect should occur because both vowels are reference vowels. In the case of /u:/ as deviant, one possible explanation for our findings could be that the additional labial feature drives the stronger effect. This additivity would then &#x0201C;overwrite&#x0201D; the feature mismatch. Several studies proved that the MMN is sensitive to an additivity effect correlated with the amount of deviating dimensions (Schr&#x000F6;ger, <xref ref-type="bibr" rid="B81">1995</xref>; Takegata et al., <xref ref-type="bibr" rid="B93">1999</xref>, <xref ref-type="bibr" rid="B94">2001a</xref>,<xref ref-type="bibr" rid="B95">b</xref>; Wolff and Schr&#x000F6;ger, <xref ref-type="bibr" rid="B102">2001</xref>). Similar observations have been made in an fMRI study in which an increasing number of features led to stronger activation in the superior temporal sulcus (STS). Besides., the effect of stronger STS activations was also seen in reaction time measures whereby reaction time decreased with increasing feature number (Scharinger et al., <xref ref-type="bibr" rid="B74">2016</xref>). Furthermore, in a MEG study, it was shown that N1m amplitudes increased when feature number increased (Scharinger et al., <xref ref-type="bibr" rid="B75">2011a</xref>). More evidence for the additive effect has been brought to light in a MEG study with consonants, in which labial, specified glides produced stronger MMFs than coronal glides (Scharinger et al., <xref ref-type="bibr" rid="B77">2011b</xref>). Under the assumption of an additivity effect of the phonological feature [LAB], we argue that the underspecification approach still holds since this model predicts effects based on sparse and abstract phonological features.</p>
<p>For the symmetrical effect in the contrast /i:/&#x02014;/a:/, there are two explanations we believe to be conceivable. The first one is that the hypothesis of NRV holds good. Since both vowels of this contrast are reference vowels in the framework, there is no discriminatory advantage in either direction. But why participants rely on phonetic features in these cases remains a question. The second more likely explanation argues within the underspecification approach: we classified Standard German /a:/ as a dorsal vowel. But there is articulatory evidence (see <xref ref-type="fig" rid="F1">Figure 1</xref>), evidence from theoretical analysis (Wiese, <xref ref-type="bibr" rid="B101">2000</xref>), and also neurobiological evidence (Obleser et al., <xref ref-type="bibr" rid="B52">2004</xref>) that Standard German /a:/ is likely not specified for a place of articulation. Thus, there is no place feature mismatch anymore for /i:/ and /a:/. Since the remaining height features [LOW] and [HIGH] are both specified and mismatching regardless of the presentation order, asymmetry has to occur.</p>
<p>In conclusion, it seems that participants use phonological and phonemic cues in vowel discrimination within natural German words. But the effects found in Experiments 1 and 2 are different although the same experimental paradigm has been applied. The reason for different effect patterns in the electrophysiological and behavioral data could lie in the different attention requirements or differences of involved processing levels between the two tasks, but to this point, it is still not clear.</p>
</sec>
</sec>
<sec id="s4">
<title>Explorative Analysis for Additional Influential Factors in Mmn and Log Rt Data</title>
<p>Because the interpretation of the MMN and RT data with common models is challenging, and because both models failed to explain the found patterns comprehensively, we decided to test for additional influential factors in both datasets.</p>
<p>Vowel perception could be influenced not only by vowel identity but also by acoustic properties like intensity, duration, and fundamental frequency (N&#x000E4;&#x000E4;t&#x000E4;nen et al., <xref ref-type="bibr" rid="B47">1989</xref>; Paavilainen et al., <xref ref-type="bibr" rid="B54">1991</xref>; Schr&#x000F6;ger, <xref ref-type="bibr" rid="B82">1996</xref>; Jacobsen et al., <xref ref-type="bibr" rid="B32">2003</xref>; Peter et al., <xref ref-type="bibr" rid="B57">2010</xref>; Pakarinen et al., <xref ref-type="bibr" rid="B55">2013</xref>). For most of these factors, researchers commonly try to exclude or control in the stimuli preparation procedure, but some acoustic factors cannot be avoided. For instance, since the phonological feature oppositions to distinguish different vowel qualities (i.e., high vowels vs. low vowels) are based on formants (Lahiri and Reetz, <xref ref-type="bibr" rid="B37">2010</xref>), they also automatically imply an acoustic difference. Moreover, when words are used as stimuli, lexical features like frequency of occurrence (Alexandrov et al., <xref ref-type="bibr" rid="B4">2011</xref>; Shtyrov et al., <xref ref-type="bibr" rid="B86">2011</xref>) or phonotactic probability (Bonte et al., <xref ref-type="bibr" rid="B8">2005</xref>; Yasin, <xref ref-type="bibr" rid="B104">2007</xref>; Emmendorfer et al., <xref ref-type="bibr" rid="B16">2020</xref>) are known to interfere in speech perception and vowel discrimination. Especially in our approach, where we tested the hypotheses of the models by using natural German spoken words, those influences may contribute to patterns of results. Therefore, even though we here focused on the identity MMN, i.e., on electrophysiological responses to physically identical stimuli in different conditions, we decided to test whether and which of these factors have an influence on our electrophysiological and which affected the behavioral data. For this purpose, we operationalized different acoustic, phonological, and lexical factors. Furthermore, we also took acoustic and perceptual factors beyond the well-known ones (e.g., degree of openness and perceived loudness) into account to disentangle their contribution to the iMMN and RT data patterns. This in-depth analysis is explorative and has never been done this extensively before.</p>
<sec id="s4-1">
<title>Preparation: Rating of Implicit Loudness</title>
<p>One possible additional influence beyond the well-known factors could be the perceived loudness of the stimuli. Here, loudness is referring to the magnitude of the auditory sensation (Fletcher and Munson, <xref ref-type="bibr" rid="B21">1933</xref>; not the physical intensity), but has been mainly taken as a perceptual correlate for sound intensity. Note that it has been shown that the physical intensity of sounds and perceived loudness are measures on different auditory dimensions. While physical intensity is stimuli-inherent, the perceived loudness of stimuli is a perceptual phenomenon and therefore subject-dependent (Yanushevskaya et al., <xref ref-type="bibr" rid="B103">2013</xref>). Moreover, while perceived loudness and sound intensity might be expected to be treated as equal, hearing research showed that two sounds of the same intensity can be rated with different perceived loudness levels due to various factors (e.g., spectral characteristics, bandwidth; Moore, <xref ref-type="bibr" rid="B46">2003</xref>). Additionally, perceived loudness levels could be correlated to gender differences since there is evidence that females perceive sounds louder than males despite the same sound pressure level (Hamamura and Iwamiya, <xref ref-type="bibr" rid="B25">2016</xref>). Furthermore, there is evidence from sound processing that cortical activations are more likely driven by perceptual factors (e.g., perceived loudness) than physical characteristics (e.g., physical intensity; Langers et al., <xref ref-type="bibr" rid="B38">2007</xref>). Therefore, it could be possible in our study, although the stimuli words were normalized for the same average intensity and vowel intensity was approximately the same across words, that participants perceived the two words in a minimal pair as strongly different in terms of perceptual or sensational loudness.</p>
<sec id="s4-1-1">
<title>Materials, Subjects, and Procedure</title>
<p>To test the word stimuli of Experiments 1 and 2 for differences in the perceived (or implicit) loudness, we conducted a rating study. Here, 10 subjects (seven females, three males) participated, all of the students or employees at Johannes Gutenberg-University of Mainz who reported normal hearing. They were all monolingual German speakers (with a mean age of 32.6 years, SD 9.6) and gave written consent before the rating.</p>
<p>The word stimuli were arranged in the same minimal pairs as in the previous experiments. We tested all five minimal pairs in both presentation orders. Because, we had five tokens per word in each presentation order, there were 25 possible combinations per presentation order, resulting in 250 trials overall. The trials were randomly arranged in ten blocks with 25 trials per block. Block order differed between subjects. The study was conducted <italic>via</italic> Presentation (version 16.4)<sup>1</sup>, and auditory stimuli were delivered <italic>via</italic> headphones on the same listening level for all participants. For the experiment, all participants were seated in a quiet room.</p>
<p>At the beginning of the experiment, the instructions were presented on the computer screen. Afterward, each trial started with a 1,500 ms blank screen. Following a fixation star to keep participants engaged with the experiment, both words were then presented (ISI: 800 ms). After the presentation of the second word, a short blank screen (600 ms) was presented before two question marks with a timeout of 2,500 ms appeared. The question marks were used as an indication for participants to give their answer <italic>via</italic> button press. Participants were instructed to rate the perceived loudness of the two words of each minimal pair in comparison to one another. Three answers were possible: first word louder, second word louder, or both words equally loud.</p>
</sec>
<sec id="s4-1-2">
<title>Analysis and Results</title>
<p>Having collected the responses of all participants, frequency values of the three answer categories (first, second, and equal) for each minimal pair per presentation order were calculated. Timeouts were not included in the analysis. The distributions of answers for each direction can be seen in <xref ref-type="fig" rid="F5">Figure 5</xref>. The Pearson chi-square test, calculated in IBM SSPS (version 21) with variables <italic>direction</italic> (10) and <italic>answer</italic> (3), showed that the relationship between both variables is highly significant (<inline-formula><mml:math id="M1"><mml:mrow><mml:msubsup><mml:mi>&#x003C7;</mml:mi><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mtext>18</mml:mtext></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mtext>2</mml:mtext></mml:msubsup></mml:mrow>
</mml:math></inline-formula> = 998.986, <italic>p</italic> &#x0003C; 0.001).</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>Results of the perceived loudness rating. The results are plotted for each presentation direction (<italic>x</italic>-axis) in relation to the frequency of the given responses (<italic>y</italic>-axis).</p></caption>
<graphic xlink:href="fnhum-15-612345-g0005.tif"/>
</fig>
<p>In preparation for the operationalization of the factor of implicit loudness for the multiple regression analysis, frequency values, transformed in percentage with the highest given answer, will be taken into account in the next step of the analysis. The percentages of answers per direction are given in <xref ref-type="table" rid="T3">Table 3</xref>.</p>
<table-wrap id="T3" position="float">
<label>Table 3</label>
<caption><p>Distribution of the answers for the perceived loudness of words (in percent) with the most given answer per presentation order in bold.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="center">Direction/condition</th>
<th align="center">Equal</th>
<th align="center">First</th>
<th align="center">Second</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">Mahl&#x02014;Mehl</td>
<td align="center"><bold>52.2</bold></td>
<td align="center">44.1</td>
<td align="center"> 3.7</td>
</tr>
<tr>
<td align="left">Mehl&#x02014;Mahl</td>
<td align="center">33.8</td>
<td align="center"> 1.3</td>
<td align="center"><bold>65.0</bold></td>
</tr>
<tr>
<td align="left">Steg&#x02014;Stieg</td>
<td align="center"><bold>70.9</bold></td>
<td align="center">12.3</td>
<td align="center">16.8</td>
</tr>
<tr>
<td align="left">Stieg&#x02014;Steg</td>
<td align="center"><bold>50.4</bold></td>
<td align="center"> 6.3</td>
<td align="center">43.3</td>
</tr>
<tr>
<td align="left">Stiel&#x02014;Stuhl</td>
<td align="center"><bold>65.7</bold></td>
<td align="center">14.9</td>
<td align="center">19.4</td>
</tr>
<tr>
<td align="left">Stuhl&#x02014;Stiel</td>
<td align="center"><bold>67.8</bold></td>
<td align="center"> 8.3</td>
<td align="center">24.0</td>
</tr>
<tr>
<td align="left">Sud&#x02014;S&#x000FC;d</td>
<td align="center">38.6</td>
<td align="center">3.8</td>
<td align="center"><bold>57.6</bold></td>
</tr>
<tr>
<td align="left">S&#x000FC;d&#x02014;Sud</td>
<td align="center"><bold>63.7</bold></td>
<td align="center">28.6</td>
<td align="center">7.8</td>
</tr>
<tr>
<td align="left">Zahl&#x02014;Ziel</td>
<td align="center">44.4</td>
<td align="center"><bold>50.6</bold></td>
<td align="center">5.0</td>
</tr>
<tr>
<td align="left">Ziel&#x02014;Zahl</td>
<td align="center">22.3</td>
<td align="center"> 1.7</td>
<td align="center"><bold>76.1</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The descriptive results are indicating a possible influence on the MMN data: in the contrast /e:/&#x02014;/a:/ (Mahl&#x02014;Mehl, Mehl&#x02014;Mahl), participants rated the word pair as equally loud by a higher percentage if <italic>Mehl</italic> was the second word (52.2%). In the reverse direction, the second presented word <italic>Mahl</italic> was more likely perceived as louder than <italic>Mehl</italic> (65%). Since, in this contrast, the MMN effect of <italic>Mahl</italic> (as deviant) was greater than in the reverse direction, implicit loudness could have a potential influence on the preattentive processing of words.</p>
<p>In the contrast /i:/&#x02014;/a:/ (Zahl&#x02014;Ziel, Ziel&#x02014;Zahl), the word <italic>Zahl</italic> was perceived more often as louder regardless of the presentation order. In the first presentation order (Zahl&#x02014;Ziel), <italic>Zahl</italic> was perceived in 50.6% as louder, and in the reverse direction, with <italic>Zahl</italic> as the second word, it was also rated as louder (76.1%). The neural data showed clear MMN effects in both directions, with an asymmetric, because stronger, the result for <italic>Zahl</italic> as deviant. It may be that the higher implicit loudness of <italic>Zahl</italic> has driven (Ziel&#x02014;Zahl) or reduced (Zahl&#x02014;Ziel) the neural effects.</p>
<p>In the next contrast with words <italic>Sud</italic> and <italic>S&#x000FC;d</italic>, similar patterns to the first one can be observed. When <italic>S&#x000FC;d</italic> was presented as the second word, participants perceived it as being louder (57.6%). When <italic>S&#x000FC;d</italic> was presented as the first word, both words were rated more often as equally loud (63.7%). Taking the MMN results into account, it might again be possible that implicit loudness affected the neural data. While the MMN effects are statistically symmetric, there is a slightly stronger effect for <italic>S&#x000FC;d</italic> as deviant (than in the reverse direction), when the plotted data are inspected.</p>
<p>In the last two contrasts (<italic>Steg</italic> and <italic>Stieg</italic>, <italic>Stuhl</italic> and <italic>Stiel</italic>), both words were described more often as equally loud within both presentation orders (Steg&#x02014;Stieg: 70.9%, Stieg&#x02014;Steg: 50.4%; Stiel&#x02014;Stuhl: 65.7%, Stuhl&#x02014;Stiel: 67.8%). Since the MMN data showed a symmetrical pattern in the statistical analysis, it could be stated that the perceived loudness might have influenced the neural effects once more.</p>
<p>Additionally, and following the feedback of participants, it can be hypothesized that implicit loudness could be correlated with the degree of openness of the long vowels. Especially with larger openness differences between vowels (/i:/&#x02014;/a:/, /e:/&#x02014;/a:/), the more open vowel /a:/ was rated as louder than the closer counterparts. Regarding the openness difference of /i:/ and /e:/, it can be stated that, phonetically, the difference here is smaller than between /i:/&#x02014;/a:/ and /e:/&#x02014;/a:/, and therefore the loudness effect could be perceptually reduced or inhibited. Moreover, the different MMN results, despite equal loudness rating patterns of /e:/&#x02014;/a:/ (asymmetric MMN) and /y:/&#x02014;/u:/ (symmetric MMN), could also support the hypothesis that the perceived loudness is correlated with the degree of openness because of the latter contrast&#x02019;s lack of height difference. Contrary, the first-mentioned contrast differs in vowel height and openness, and therefore the influence of perceived loudness could lead to stronger neural effects.</p>
</sec>
</sec>
<sec id="s4-2">
<title>Explorative Analysis <italic>via</italic> Multiple Regressions</title>
<p>Because of the challenging and unexpected electrophysiological effects that do not match the behavioral results in addition to the descriptive identification of a potential influence of the implicit loudness, we decided to also investigate the possible influences of several additional factors on the neural (iMMN difference values of mean voltages between standard and deviant of the same stimulus) as well as the behavioral level (log RTs).</p>
<sec id="s4-2-1">
<title>Defining Factors</title>
<p>Fifteen potential influential factors were defined, based on theoretical (<italic>specificity, peripherality, focality</italic>) and empirical input (<italic>implicit loudness, electrodes, degree of openness</italic>) as well as stimulus-inherent characteristics (<italic>F1, F2, F3, frequency of words, bigram frequency, f0, vowel duration, intensity</italic>). Additionally, one control factor (<italic>contrast</italic>) has been taken into account. All factors were operationalized and calculated with mean values per word category (i.e., mean F1 for <italic>Mahl</italic>) since the iMMN data as well as log RT data were also obtained as averaged data (for example in iMMN data, all tokens for <italic>Mahl</italic> as deviant or standard are collapsed in respectively one mean amplitude value).</p>
<p>The theoretical factors of <italic>specificity</italic> and <italic>peripherality</italic> have been operationalized concerning both evaluated models in Experiments 1 and 2. While <italic>specificity</italic> (difference value between the number abstract features in the deviant minus the number of features in the standard) refers to FUL and takes the additivity effect discussed in the RT data into account, <italic>peripherality</italic> has been operationalized according to the assumption of NRV that peripheral vowels may be more salient and act as referents in vowel discrimination as a categorical variable (discrimination towards a more peripheral vowel, towards a more central vowel, or no referent/equal position in the vowel space). Additionally, <italic>focality</italic> has been operationalized according to the notion in NRV that the universal preference of referent vowels may be alternated in adults due to language experience and is therefore taking the formant convergences in our stimulus set into account. Focality was again operationalized as a categorical variable (discrimination from a less to a more focal vowel or from a more focal to a less focal vowel).</p>
<p>The empirically motivating factors have been chosen from the input of the loudness rating. As mentioned before, the results of the rating study suggest an influence of <italic>implicit loudness</italic> on the neural data. Therefore, this factor has been included in the further analysis as a function of the answer category with the highest percentage per presentation order (first word louder, second word louder, equally loud). Because there could be a possible relationship between loudness and the openness of vowels, the <italic>degree of openness</italic> (increasing openness of the mouth, decreasing openness, equal openness) was also taken into account.</p>
<p>Since, we tested natural (and mostly unmanipulated) German words in this study, there were stimuli-inherent differences between words that were controlled but could not be excluded in the preparation of the stimuli. The three first factors of this category are the differences between standard and deviant in terms of a change in <italic>F1, F2, and F3</italic>. To display the presentation orders of the stimuli in this factor, difference values (e.g., mean F1 of deviant minus mean F1 of standard) were calculated. Another possible stimulus-inherent influence is the <italic>frequency</italic> of occurrence of the words used. Since, we wanted to test a large set of long vowels, we had to choose monosyllabic minimal pairs to reduce testing time. Being restricted by the German lexicon, we were not able to perfectly balance the lexical frequency of words; therefore, there are frequency differences between the words making up the minimal pairs. To test the influence of the <italic>frequency</italic> of occurrence on the electrophysiological and behavioral patterns, we included this factor in the explorative analysis in the form of difference values (log lexical frequency of the deviant minus log lexical frequency standard). The same was true for the factor <italic>bigram frequency</italic> (bigram frequency deviant minus bigram frequency standard). The same holds good for the factors <italic>f0, vowel duration</italic>, and <italic>intensity</italic>. To evaluate the influence of the stimuli-inherent differences in fundamental frequency, vowel duration, and intensity, we operationalized those factors also as difference values between deviant and standard (<italic>f0</italic>: the difference between mean f0 of the deviant minus mean f0 of the standard; <italic>vowel duration</italic>: the difference between vowel duration of the deviant minus the vowel duration of the standard; <italic>intensity</italic>: the difference between mean vowel intensity of the deviant minus mean vowel intensity of the standard).</p>
<p>Last but not least, the <italic>contrast</italic> has been included as a controlling factor, since the iMMN and RT effects were different for the tested vowel oppositions. Here, both presentation orders of each minimal pair are combined.</p>
</sec>
<sec id="s4-2-2">
<title>Correlation and Single Linear Regressions</title>
<p>In preparation for the multiple regression analysis, we conducted first Kendall&#x02019;s Tau correlation for all previously defined factors. Additionally, we calculated for each factor a single regression model on the MMN data to identify reasonable factors to be included in the final multiple regression analysis.</p>
<p>Correlation analysis showed very strong correlation between the factors <italic>F2</italic> and <italic>specificity</italic> (&#x003C4; = &#x02212;0.907, <italic>p</italic> &#x0003C; 0.001) and <italic>F1</italic> and <italic>degree of openness</italic> (&#x003C4; = 0.866, <italic>p</italic> &#x0003C; 0.001), <italic>vowel duration</italic> and <italic>degree of openness</italic> (&#x003C4; = 0.856, <italic>p</italic> &#x0003C; 0.001), as well as <italic>F2</italic> and <italic>F3</italic> (&#x003C4; = 0.867, <italic>p</italic> &#x0003C; 0.001). Because of that, we only included <italic>specificity</italic>, <italic>F1</italic>, and <italic>F3</italic> as theoretically implied factors. Additionally, <italic>vowel</italic> <italic>duration</italic> was included in the analysis since there is evidence that sound duration is influencing the perceived loudness (Todd and Michie, <xref ref-type="bibr" rid="B97">2000</xref>). The exclusion of the other factors was necessary to avoid collinearities.</p>
<p>Additionally, strong, but in respect with collinearity uncritical correlations, were found between the following factors: <italic>F1</italic> and <italic>f0</italic> (&#x003C4; = &#x02212;0.764, <italic>p</italic> &#x0003C; 0.001), <italic>F3</italic> and <italic>specificity</italic> (&#x003C4; = &#x02212;0.760, <italic>p</italic> &#x0003C; 0.001), <italic>F1</italic> and <italic>vowel duration</italic> (&#x003C4; = 0.764, <italic>p</italic> &#x0003C; 0.001), <italic>vowel duration</italic> and <italic>f0</italic> (&#x003C4; = &#x02212;0.778, <italic>p</italic> &#x0003C; 0.001), <italic>specificity</italic> and <italic>peripherality</italic> (&#x003C4; = 0.559, <italic>p</italic> &#x0003C; 0.001), <italic>bigram frequency</italic> and <italic>peripherality</italic> (&#x003C4; = 0.676, <italic>p</italic> &#x0003C; 0.001), <italic>vowel duration</italic> and <italic>intensity</italic> (&#x003C4; = &#x02212;0.689, <italic>p</italic> &#x0003C; 0.001), <italic>f0</italic> and <italic>intensity</italic> (&#x003C4; = 0.556, <italic>p</italic> &#x0003C; 0.001), <italic>focality</italic> and <italic>F1</italic> (&#x003C4; = &#x02212;0.603, <italic>p</italic> &#x0003C; 0.001), <italic>F2</italic> (&#x003C4; = 0.507, <italic>p</italic> &#x0003C; 0.001) as well as F3 (&#x003C4; = 0.686, <italic>p</italic> &#x0003C; 0.001), <italic>focality</italic> and <italic>f0</italic> (&#x003C4; = 0.686, <italic>p</italic> &#x0003C; 0.001), <italic>focality</italic> and <italic>vowel duration</italic> (&#x003C4; = &#x02212;0.745, <italic>p</italic> &#x0003C; 0.001), with <italic>intensity</italic> (&#x003C4; = 0.566, <italic>p</italic> &#x0003C; 0.001), with <italic>degree of openness</italic> (&#x003C4; = &#x02212;0.731, <italic>p</italic> &#x0003C; 0.001) and <italic>implicit</italic> <italic>loudness</italic> (&#x003C4; = &#x02212;0.654, <italic>p</italic> &#x0003C; 0.001), <italic>implicit loudness</italic> and <italic>f0</italic> (&#x003C4; = &#x02212;0.775, <italic>p</italic> &#x0003C; 0.001), <italic>implicit loudness</italic> and <italic>F1</italic> (&#x003C4; = 0.667, <italic>p</italic> &#x0003C; 0.001), <italic>implicit loudness</italic> and <italic>vowel duration</italic> (&#x003C4; = 0.660, <italic>p</italic> &#x0003C; 0.001), <italic>implicit loudness</italic> and <italic>degree of openness</italic> (&#x003C4; = 0.603, <italic>p</italic> &#x0003C; 0.001), and <italic>implicit loudness</italic> and <italic>intensity</italic> (&#x003C4; = &#x02212;0.488, <italic>p</italic> &#x0003C; 0.001).</p>
<p>Single factor regression (with MMN data) revealed the influence of only five significant factors with reasonable <italic>R</italic><sup>2</sup> and adjusted <italic>R</italic><sup>2</sup> values: <italic>implicit loudness</italic> (&#x00394;<italic>R</italic><sup>2</sup> = 0.023, <italic>p</italic> &#x0003C; 0.001), <italic>contrast</italic> (&#x00394;<italic>R</italic><sup>2</sup> = 0.019, <italic>p</italic> &#x0003C; 0.001), <italic>intensity</italic> (&#x00394;<italic>R</italic><sup>2</sup> = 0.016, <italic>p</italic> &#x0003C; 0.001), <italic>f0</italic> (&#x00394;<italic>R</italic><sup>2</sup> = 0.012, <italic>p</italic> &#x02264; 0.001), <italic>vowel duration</italic> (&#x00394;<italic>R</italic><sup>2</sup> = 0.013, <italic>p</italic> &#x0003C; 0.001).</p>
</sec>
<sec id="s4-2-3">
<title>Hierarchical Multiple Regressions for MMN and Log RT Data</title>
<p>The five previously identified factors were applied to hierarchical multiple regression with separate calculations for the MMN and the RT datasets in five steps with the following order: implicit loudness (model 1), contrast (model 2), f0 (model 3), vowel duration (model 4), and intensity (Model 5).</p>
<p>Results for the MMN dataset, for which the relevant key figures are displayed in <xref ref-type="table" rid="T4">Table 4</xref>, indicate that only the first two factors (implicit loudness and vowel contrast) are contributing to account for variance and that implicit loudness and contrast are approximately an equal fit for the explanation of results. In this study, implicit loudness seems to influence the neural results of the MMN study (see <xref ref-type="fig" rid="F6">Figure 6</xref>).</p>
<table-wrap id="T4" position="float">
<label>Table 4</label>
<caption><p>Multiple regression models for the MMN dataset.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="center">Model&#x00023;</th>
<th align="center"></th>
<th align="center">b</th>
<th align="center">SE B</th>
<th align="center">&#x003B2;</th>
<th align="center"><italic>p</italic></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">1</td>
<td align="center">Constant</td>
<td align="center">&#x02212;0.363 (&#x02212;0.592, &#x02212;0.133)</td>
<td align="center">0.117</td>
<td/>
<td align="center">0.002</td>
</tr>
<tr>
<td/>
<td align="center">Loudness</td>
<td align="center">&#x02212;0.436 (&#x02212;0.607, &#x02212;0.265)</td>
<td align="center">0.087</td>
<td align="center">&#x02212;0.155</td>
<td align="center">0.000</td>
</tr>
<tr>
<td align="left"><italic>R</italic><sup>2</sup> = 0.024, &#x00394;<italic>R</italic><sup>2</sup> = 0.023, <italic>p</italic> &#x0003C; 0.001</td>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">2</td>
<td align="center">Constant</td>
<td align="center">&#x02212;0.703 (&#x02212;0.971, &#x02212;0.434)</td>
<td align="center">0.137</td>
<td/>
<td align="center">0.000</td>
</tr>
<tr>
<td/>
<td align="center">Loudness</td>
<td align="center">&#x02212;0.436 (&#x02212;0.605, &#x02212;0.267)</td>
<td align="center">0.086</td>
<td align="center">&#x02212;0.155</td>
<td align="center">0.002</td>
</tr>
<tr>
<td/>
<td align="center">Contrast</td>
<td align="center">0.170 (0.098, 0.242)</td>
<td align="center">0.037</td>
<td align="center">0.143</td>
<td align="center">0.000</td>
</tr>
<tr>
<td align="left"><italic>R</italic><sup>2</sup> = 0.044, &#x00394;<italic>R</italic><sup>2</sup> = 0.043, <italic>p</italic> &#x0003C; 0.001</td>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>Scatterplot for the regression analysis of the given iMMN data from Experiment 1. MMN difference values of each participant (<italic>y</italic>-axis) per vowel contrast (<italic>x</italic>-axis) in relation to implicit loudness. Increasing loudness (deviant louder than standard) is shown as a blue triangle, decreasing loudness (standard louder than deviant) as a red triangle, and equal perceived loudness as a green dot. MMN difference values are scaled with perceiving the stimuli as equally loud (most clearly seen in vowel contrasts StiegSteg and StielStuhl).</p></caption>
<graphic xlink:href="fnhum-15-612345-g0006.tif"/>
</fig>
<p>In contrast, results of the second multiple regression model in the log RT dataset implicate that the factors influencing the neural results are not contributing to the explanation of behavioral patterns. Once again, Model 2 (implicit loudness and contrast) is fitting best with contrast as the only significant contributing regression coefficient (<xref ref-type="table" rid="T5">Table 5</xref>). Thus, implicit loudness is not contributing to the found behavioral pattern in the reaction time experiment (see <xref ref-type="fig" rid="F7">Figure 7</xref>).</p>
<table-wrap id="T5" position="float">
<label>Table 5</label>
<caption><p>Multiple regression models for the reaction time (RT) dataset (n.s. = no significance).</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="center">Model&#x00023;</th>
<th align="center"></th>
<th align="center" colspan="1">b</th>
<th align="center">SE B</th>
<th align="center">&#x003B2;</th>
<th align="center"><italic>p</italic></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">1</td>
<td align="center">Constant</td>
<td align="center">2.547 (2.522,2.573)</td>
<td align="center">0.013</td>
<td/>
<td align="center">0.000</td>
</tr>
<tr>
<td/>
<td align="center">Loudness</td>
<td align="center">0.001 (&#x02212;0.021,0.024)</td>
<td align="center">0.011</td>
<td align="center">0.010</td>
<td align="center">n.s.</td>
</tr>
<tr>
<td align="left"><italic>R</italic><sup>2</sup> = 0.000, &#x00394;<italic>R</italic><sup>2</sup> = &#x02212;0.006, n.s.</td>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
<tr>
<td align="left">2</td>
<td align="center">Constant</td>
<td align="center">2.558 (2.535,2.595)</td>
<td align="center">0.015</td>
<td/>
<td align="center">0.000</td>
</tr>
<tr>
<td/>
<td align="center">Loudness</td>
<td align="center">0.001 (&#x02212;0.020,0.023)</td>
<td align="center">0.011</td>
<td align="center">0.127</td>
<td align="center">n.s.</td>
</tr>
<tr>
<td/>
<td align="center">Contrast</td>
<td align="center">&#x02212;0.021 (&#x02212;0.033&#x02032;6, &#x02212;0.005)</td>
<td align="center">0.008</td>
<td align="center">&#x02212;0.204</td>
<td align="center">0.011</td>
</tr>
<tr>
<td align="left"><italic>R</italic><sup>2</sup> = 0.041, &#x00394;<italic>R</italic><sup>2</sup> = 0.029, <italic>p</italic> &#x0003C; 0.05</td>
<td/>
<td/>
<td/>
<td/>
<td/>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>Scatterplot for the regression of the obtained reaction time (RT) data from Experiment 2. Mean log RTs of each subject (<italic>y</italic>-axis) are depicted per vowel contrast (<italic>x</italic>-axis) in relation to implicit loudness. RT results (log-values) are not scaled by the perceived loudness of the stimuli.</p></caption>
<graphic xlink:href="fnhum-15-612345-g0007.tif"/>
</fig>
</sec>
</sec>
<sec sec-type="discussion" id="s4-3">
<title>Discussion</title>
<p>Multiple regression analysis on both datasets revealed that the perceptual factor of perceived loudness had only an influence on neural effects. Here, it can be stated that regarding the MMN data, the phonological and phonetic status of the vowels presented in the minimal pairs played a role in the elicitation of MMN effects (factor contrast), but&#x02014;crucially&#x02014;effects were simultaneously driven by the implicit loudness of the presented words. Overall, it seems that the neural effect was scaled with perceiving the stimuli as equally loud. Put differently: the more strongly one word was perceived as being louder, the further the MMN difference values deviated from zero. Therefore, it can be argued that perceived, or implicit, loudness seems to be an important factor in interpreting the found neural patterns.</p>
<p>This is especially true for those contrasts for which both models (NRV and FUL) failed to explain the effects. Especially in the symmetrical contrasts of <italic>Stieg&#x02014;Steg</italic> and <italic>Stiel&#x02014;Stuhl</italic>, implicit loudness seems to drive the symmetry as in this contrast, both words were more often perceived as equally loud regardless of the presentation order. The small visible (but statistically not significant effect) in the contrast <italic>S&#x000FC;d&#x02014;Sud</italic> could also be explained by this factor since there were more increasing judgments if <italic>S&#x000FC;d</italic> was the second word. Missing statistical significance could be a result of the correlation of loudness with a degree of openness. Since this contrast did not differ in terms of vowel height (and therefore openness), the influence of perceived loudness could be weaker than in vowel oppositions with height differences. Turning to the asymmetrical patterns in the MMN data, it can be stated that for the pattern of <italic>Ziel&#x02014;Zahl</italic>, which was especially challenging because both models could not explain the found asymmetry, implicit loudness once more seems to drive the neural effects since <italic>Zahl</italic> was more often perceived as louder regardless of the presentation order. If <italic>Zahl</italic> served as deviant, the greater perceived loudness has led to a stronger effect than in the reverse direction. Here, the greater implicit loudness of the standard could have reduced MMN effects. In the last contrast (i.e., <italic>Mehl&#x02014;Mahl)</italic>, differences in perceived loudness and degree of openness could have led to the stronger effect for <italic>Mahl</italic> (as deviant). <italic>Mahl</italic> as the second presented word was perceived as louder than in the reverse direction. In this direction, the degree of openness is also increasing. In the reverse direction, equally perceived loudness might have elicited smaller effects.</p>
<p>Turning to the RT data, multiple regression analysis revealed that the perceived loudness did not influence the behavioral patterns; therefore, the effects are more likely driven by the traditional models (namely FUL and NRV) discussed before.</p>
</sec>
</sec>
<sec id="s5">
<title>General Discussion and Conclusion</title>
<p>In this article, we reported the results of an electrophysiological and a behavioral study as well as an explorative analysis of influential factors on vowel discrimination. While most studies investigating speech sound discrimination only tested two or three vowel oppositions, we conducted our MMN study on a much larger stimulus set. Here, we investigated preattentive vowel processing with five different vowel contrasts covering the most important German long vowels embedded in real and natural German words. To obtain an even more natural listening situation, we used five tokens per word, which resulted in a large stimulus set. The purpose of the investigations reported here was 2-fold: first, we wanted to compare two often discussed models for vowel discrimination to investigate which model can explain the found effects in German in the best way. Second, we wanted to shed further light on factors influencing vowel discrimination on the neural and behavioral levels. For this purpose, we conducted an in-depth analysis delving into possible confounds to a degree that has not been investigated so far.</p>
<p>To summarize the results of the electrophysiological experiment concerning the first research question, we found MMN evidence for discrimination and perceptual asymmetries (or symmetries) in vowel perception according to the NRV model. Three contrasts showed facilitated and asymmetric discrimination on presenting a less peripheral vowel as standard and a more peripheral vowel as deviant. These results are in line with other behavioral and electrophysiological studies (e.g., Masapollo et al., <xref ref-type="bibr" rid="B41">2015</xref>, <xref ref-type="bibr" rid="B44">2017b</xref>; Zhao et al., <xref ref-type="bibr" rid="B105">2019</xref>) that also report easier discrimination from a more central to a more peripheral vowel. Only one contrast could be explained within the underspecification approach. By contrast, both models failed to explain the found symmetric neural patterns. For those vowel oppositions, it could be possible that the phonemic discrimination was overridden in preattentive processing through acoustic proximity (Pettigrew et al., <xref ref-type="bibr" rid="B59">2004b</xref>) or sensational interferences caused by perceived loudness. The lack of phonemic discrimination and weighting of sensational influences in the MMN experiment could be due to experimental protocol since the subjects were not instructed to perform phonemic discrimination, but to ignore the stimulation (Johnson, <xref ref-type="bibr" rid="B33">2015</xref>). To test those challenging results, we investigated these contrasts with a behavioral active oddball paradigm and instructed participants here to perform a phonemic decision. Thus, we assumed that in the active oddball paradigm subjects had to activate more abstract mental representation more strongly due to allophonic variance in the stimuli; therefore blending out simple acoustic differences in the decision making. The behavioral results, in contrast to the MMN effects, showed that participants were able of phonemic discrimination based on abstract representations. Here, the found patterns can only be fully explained by the underspecification approach, in line with previous studies delivering evidence for speech sound discrimination with the help of sparse and abstract features (e.g., Eulitz and Lahiri, <xref ref-type="bibr" rid="B20">2004</xref>; Lahiri and Reetz, <xref ref-type="bibr" rid="B37">2010</xref>; Scharinger et al., <xref ref-type="bibr" rid="B73">2012a</xref>, <xref ref-type="bibr" rid="B78">b</xref>).</p>
<p>In summary, the results of Experiments 1 and 2 are challenging in two ways. First, the neural pattern cannot be explained comprehensively by either of the two models. Second, the neural and behavioral patterns do not match. The lack of compliance between electrophysiological and behavioral results can be interpreted in terms of an attention shift and cue weighting as a function of task dependency. Differences in cue weighting due to attention shifts have been reported in several studies. Szymanski et al. (<xref ref-type="bibr" rid="B91">1999</xref>) conducted an MMN study with and without attention on the stimulation and found differences in the neural responses. They interpreted these findings in terms of a modulation of the memory trace in the attended condition. Here, attention leads to the activation of more accurate and precise representations of the standards, which in turn generate larger responses of the deviant. Therefore, it can be argued that in attended stimulation, more information is accessible for discrimination than in unattended conditions. Similar results concerning the richness of mental representation accessible during discrimination as a function of attention were also found by Tuomainen et al. (<xref ref-type="bibr" rid="B98">2013</xref>). In an MMN study with an active go/no-go task, they found that in the attentive task participants were able to use more spectral attributes in vowel discrimination than when they listened passively to the stimulation. The authors interpreted the results as a change in perceptual discrimination strategy due to the attention shift. Furthermore, Savela et al. (<xref ref-type="bibr" rid="B71">2003</xref>) found, in a study combining MMN (passive oddball) and RT (active oddball), that subjects discriminated the used Finnish and Komi vowels differently depending on the task. While the behavioral results indicated phonemic discrimination of the vowels, the preattentive MMN patterns were more driven by acoustic differences than phonemic representations. Concerning our electrophysiological and behavioral results, it can be assumed that the attention shift between the passive and the active oddball has led to differences in cue weighting in vowel discrimination. We argue that in the active experiment, participants were able to discriminate phonemically, which led to patterns explainable by common models. In contrast to this&#x02014;but following previous studies&#x02014;the passive MMN patterns are based not only on phonemic but also on acoustic or perceptual differences.</p>
<p>To address this issue further, we conducted an explorative analysis of influencing factors of the electrophysiological as well the behavioral datasets. We included several theoretical, lexical, and phonotactic factors that are known to influence results. While other studies found an influence on neural data, for example, of phonotactic probabilities (Bonte et al., <xref ref-type="bibr" rid="B8">2005</xref>; Yasin, <xref ref-type="bibr" rid="B104">2007</xref>; Emmendorfer et al., <xref ref-type="bibr" rid="B16">2020</xref>) or the lexical frequency of words (Alexandrov et al., <xref ref-type="bibr" rid="B4">2011</xref>; Shtyrov et al., <xref ref-type="bibr" rid="B86">2011</xref>; Aleksandrov et al., <xref ref-type="bibr" rid="B3">2017</xref>), we cannot provide evidence for those factors neither on the electrophysiological nor on the behavioral data. On the contrary, we have identified a new influencing factor on MMN data: we found that neural effects were not only driven by phonemic features but also by the perceptual and psychoacoustic differences in perceived loudness in the stimuli. In contrast, no such influence of this factor could be found in the behavioral data. Therefore, the multiple regression analyses on both datasets support the aforementioned interpretation on different discrimination strategies since we found that the influence of perceived loudness of the word stimuli only mattered in the neural but not the behavioral data. Once again, these results can be interpreted as evidence that in preattentive processing, more perceptual and acoustic features are responsible for the elicitation of effects. But when attention was shifted towards the stimulation (like in the active oddball paradigm of the RT experiment), these perceptual factors receded into the background, and discrimination was based on phonemic representations of the perceived vowels only.</p>
<p>Although perceived loudness is related to (and heavily determined by) sound intensity, two sounds of equal perceived loudness may well have different levels of sound intensity (Yanushevskaya et al., <xref ref-type="bibr" rid="B103">2013</xref>). This is due to the processing of auditory stimuli in the cochlea (Moore, <xref ref-type="bibr" rid="B46">2003</xref>), which depends not only on the characteristics of the stimuli, such as bandwidth but on the listener as well. We found evidence that speech signals of approximately equal intensity could still be perceived to be of different loudness. Additionally, we could show that within our datasets, perceived loudness was highly positively correlated with the degree of openness of the vowels and changes in F1. We conclude that perceived loudness differences could be guided by differences in the degree of openness since with increasing openness of the tested vowels, the perceived loudness of words increased as well, and since increasing loudness elicited larger MMN effects. These results add evidence to the hypothesis that perceived loudness of vowel stimuli is also linked to vowel quality (Glave and Rietveld, <xref ref-type="bibr" rid="B23">1975</xref>, <xref ref-type="bibr" rid="B24">1979</xref>). Additionally, we found correlations of perceived loudness and changes in f0, intensity, and vowel duration. But the multiple regression analysis showed that those additional factors did not contribute to the found neural asymmetries. Here, only the differences in perceived loudness can explain the found patterns. However, since there is evidence that perceived loudness can be influenced by vowel duration (Todd and Michie, <xref ref-type="bibr" rid="B97">2000</xref>) and changes in fundamental frequency (Hsu et al., <xref ref-type="bibr" rid="B30">2015</xref>), more studies are needed to disentangle all the factors contributing to differences in the sensational perceived loudness of stimuli and influencing natural vowel processing.</p>
<p>To our knowledge, we are the first to find evidence for the influence of perceived loudness on the perception of German long vowels and MMN data regarding natural vowel processing. We propose that the perceptual, or implicit, the loudness of stimuli can act as an intermediate representation level between stimuli-inherent acoustics and abstract phonological features. The exact influence of perceptual and psychoacoustic factors, like perceived loudness in speech processing, is still underinvestigated and more research is needed. But for the time being, our results provide evidence that studies should include more factors beyond the well-known (and theoretically driven) when analyzing and interpreting neural and behavioral data.</p>
</sec>
<sec id="s6">
<title>Data Availability Statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec id="s7">
<title>Ethics Statement</title>
<p>The studies involving human participants were reviewed and approved by Ethics Committee of the Society of German Linguistics (DGfS). The patients/participants provided their written informed consent to participate in this study.</p>
</sec>
<sec id="s8">
<title>Author Contributions</title>
<p>MR contributed to the design of the work, acquisition, analysis, and interpretation of the data, as well as drafting of the work. AW contributed to the design of the work, acquisition, and revising the manuscript. AN contributed to the design of the work, acquisition and analysis of the data, as well as the revising of the manuscript. MS contributed to the design of the work, analysis and interpretation of the data, as well as drafting and revising of the manuscript. All authors gave final approval of the version to be submitted and agreed to be accountable for all aspects of the work ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved. All authors contributed to the article and approved the submitted version.</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</body>
<back>
<fn-group>
<fn fn-type="financial-disclosure">
<p><bold>Funding.</bold> The research presented here was supported by the Hessen State Ministry of Higher Education, Research and the Arts, Landesoffensive Zur Entwicklung Wissenschaftliche-&#x000F6;konomischer Excellenz (LOEWE), Research Focus: exploring fundamental linguistic categories grant, Research Project: phonological word&#x02014;Constituents of the Phonological Word, and the Johannes Gutenberg University Mainz.</p>
</fn>
</fn-group>
<sec id="s10">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fnhum.2021.612345/full&#x00023;supplementary-material">https://www.frontiersin.org/articles/10.3389/fnhum.2021.612345/full&#x00023;supplementary-material</ext-link>.</p>
<supplementary-material xlink:href="Table_1.XLSX" id="SM1" mimetype="application/xlsx" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Aaltonen</surname> <given-names>O.</given-names></name> <name><surname>Eerola</surname> <given-names>O.</given-names></name> <name><surname>Lang</surname> <given-names>A. H.</given-names></name> <name><surname>Uusipaikka</surname> <given-names>E.</given-names></name> <name><surname>Tuomainen</surname> <given-names>J.</given-names></name></person-group> (<year>1994</year>). <article-title>Automatic discrimination of phonetically relevant and irrelevant vowel parameters as reflected by mismatch negativity</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>96</volume>, <fpage>1489</fpage>&#x02013;<lpage>1493</lpage>. <pub-id pub-id-type="doi">10.1121/1.410291</pub-id><pub-id pub-id-type="pmid">7963013</pub-id></citation></ref>
<ref id="B2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Aichert</surname> <given-names>I.</given-names></name> <name><surname>Marquardt</surname> <given-names>C.</given-names></name> <name><surname>Ziegler</surname> <given-names>W.</given-names></name></person-group> (<year>2005</year>). <article-title>Frequenzen sublexikalischer Einheiten des Deutschen: CELEX-basierte Datenbanken</article-title>. <source>Neurolinguistik</source> <volume>19</volume>, <fpage>55</fpage>&#x02013;<lpage>81</lpage>.</citation></ref>
<ref id="B3"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Aleksandrov</surname> <given-names>A. A.</given-names></name> <name><surname>Memetova</surname> <given-names>K. S.</given-names></name> <name><surname>Stankevich</surname> <given-names>L. N.</given-names></name> <name><surname>Uplisova</surname> <given-names>K. O.</given-names></name></person-group> (<year>2017</year>). <article-title>Effects of Russian-language word frequency on mismatch negativity in auditory event-related potentials</article-title>. <source>Neurosci. Behav. Phys.</source> <volume>47</volume>, <fpage>1043</fpage>&#x02013;<lpage>1050</lpage>. <pub-id pub-id-type="doi">10.1007/s11055-017-0510-3</pub-id><pub-id pub-id-type="pmid">30193042</pub-id></citation></ref>
<ref id="B4"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alexandrov</surname> <given-names>A. A.</given-names></name> <name><surname>Boricheva</surname> <given-names>D. O.</given-names></name> <name><surname>Pulverm&#x000FC;ller</surname> <given-names>F.</given-names></name> <name><surname>Shtyrov</surname> <given-names>Y.</given-names></name></person-group> (<year>2011</year>). <article-title>Strength of word-specific neural memory traces assessed electrophysiologically</article-title>. <source>PLoS One</source> <volume>6</volume>:<fpage>e22999</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0022999</pub-id><pub-id pub-id-type="pmid">21853063</pub-id></citation></ref>
<ref id="B5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Archangeli</surname> <given-names>D.</given-names></name></person-group> (<year>1988</year>). <article-title>Aspects of underspecification theory</article-title>. <source>Phonology</source> <volume>5</volume>, <fpage>183</fpage>&#x02013;<lpage>207</lpage>. <pub-id pub-id-type="doi">10.1017/S0952675700002268</pub-id></citation></ref>
<ref id="B6"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Boersma</surname> <given-names>P.</given-names></name> <name><surname>Weenink</surname> <given-names>D.</given-names></name></person-group> (<year>2016</year>). <source>Praat: Doing Phonetics by Computer (6.0.18) [Computer Software]</source>. <publisher-name>Phonetic Science</publisher-name>.</citation></ref>
<ref id="B7"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bohn</surname> <given-names>O.-S.</given-names></name> <name><surname>Polka</surname> <given-names>L.</given-names></name></person-group> (<year>2001</year>). <article-title>Target spectral, dynamic spectral and duration cues in infant perception of German vowels</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>110</volume>, <fpage>504</fpage>&#x02013;<lpage>515</lpage>. <pub-id pub-id-type="doi">10.1121/1.1380415</pub-id><pub-id pub-id-type="pmid">11508975</pub-id></citation></ref>
<ref id="B8"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bonte</surname> <given-names>M.</given-names></name> <name><surname>Mitterer</surname> <given-names>H.</given-names></name> <name><surname>Zellagui</surname> <given-names>N.</given-names></name> <name><surname>Poelmans</surname> <given-names>H.</given-names></name> <name><surname>Blomert</surname> <given-names>L.</given-names></name></person-group> (<year>2005</year>). <article-title>Auditory cortical tuning to statistical regularities in phonology</article-title>. <source>Clin. Neurophysiol.</source> <volume>116</volume>, <fpage>2765</fpage>&#x02013;<lpage>2774</lpage>. <pub-id pub-id-type="doi">10.1016/j.clinph.2005.08.012</pub-id><pub-id pub-id-type="pmid">16256430</pub-id></citation></ref>
<ref id="B9"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brysbaert</surname> <given-names>M.</given-names></name> <name><surname>Buchmeier</surname> <given-names>M.</given-names></name> <name><surname>Conrad</surname> <given-names>M.</given-names></name> <name><surname>Jacobs</surname> <given-names>A. M.</given-names></name> <name><surname>B&#x000F6;lte</surname> <given-names>J.</given-names></name> <name><surname>B&#x000F6;hl</surname> <given-names>A.</given-names></name></person-group> (<year>2011</year>). <article-title>The word frequency effect: a review of recent developments and implications for the choice of frequency estimates in German</article-title>. <source>Exp. Psychol.</source> <volume>58</volume>, <fpage>412</fpage>&#x02013;<lpage>424</lpage>. <pub-id pub-id-type="doi">10.1027/1618-3169/a000123</pub-id><pub-id pub-id-type="pmid">21768069</pub-id></citation></ref>
<ref id="B10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cornell</surname> <given-names>S. A.</given-names></name> <name><surname>Lahiri</surname> <given-names>A.</given-names></name> <name><surname>Eulitz</surname> <given-names>C.</given-names></name></person-group> (<year>2013</year>). <article-title>Inequality across consonantal contrasts in speech perception: evidence from mismatch negativity</article-title>. <source>J. Exp. Psychol. Hum. Percept. Perform.</source> <volume>39</volume>, <fpage>757</fpage>&#x02013;<lpage>772</lpage>. <pub-id pub-id-type="doi">10.1037/a0030862</pub-id><pub-id pub-id-type="pmid">23276108</pub-id></citation></ref>
<ref id="B11"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cummings</surname> <given-names>A.</given-names></name> <name><surname>Madden</surname> <given-names>J.</given-names></name> <name><surname>Hefta</surname> <given-names>K.</given-names></name></person-group> (<year>2017</year>). <article-title>Converging evidence for [coronal] underspecification in English-speaking adults</article-title>. <source>J. Neurolinguistics</source> <volume>44</volume>, <fpage>147</fpage>&#x02013;<lpage>162</lpage>. <pub-id pub-id-type="doi">10.1016/j.jneuroling.2017.05.003</pub-id><pub-id pub-id-type="pmid">29085183</pub-id></citation></ref>
<ref id="B12"><citation citation-type="book"><person-group person-group-type="author"><name><surname>de Jonge</surname> <given-names>M. J. I.</given-names></name> <name><surname>Boersma</surname> <given-names>P.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;French high-mid vowels are underspecified for heigh</article-title>,&#x0201D; in <source>Proceedings of the 18th International Congress of Phonetic Sciences</source> <volume>5</volume> (<publisher-loc>Glasgow</publisher-loc>: <publisher-name>The University of Glasgow</publisher-name>).</citation></ref>
<ref id="B13"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Deguchi</surname> <given-names>C.</given-names></name> <name><surname>Chobert</surname> <given-names>J.</given-names></name> <name><surname>Brunelli&#x000E8;re</surname> <given-names>A.</given-names></name> <name><surname>Nguyen</surname> <given-names>N.</given-names></name> <name><surname>Colombo</surname> <given-names>L.</given-names></name> <name><surname>Besson</surname> <given-names>M.</given-names></name></person-group> (<year>2010</year>). <article-title>Pre-attentive and attentive processing of french vowels</article-title>. <source>Brain Res.</source> <volume>1366</volume>, <fpage>149</fpage>&#x02013;<lpage>161</lpage>. <pub-id pub-id-type="doi">10.1016/j.brainres.2010.09.104</pub-id><pub-id pub-id-type="pmid">20920484</pub-id></citation></ref>
<ref id="B14"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dehaene-Lambertz</surname> <given-names>G.</given-names></name> <name><surname>Dupoux</surname> <given-names>E.</given-names></name> <name><surname>Gout</surname> <given-names>A.</given-names></name></person-group> (<year>2000</year>). <article-title>Electrophysiological correlates of phonological processing: a cross-linguistic study</article-title>. <source>J. Cogn. Neurosci.</source> <volume>12</volume>, <fpage>635</fpage>&#x02013;<lpage>647</lpage>. <pub-id pub-id-type="doi">10.1162/089892900562390</pub-id><pub-id pub-id-type="pmid">10936916</pub-id></citation></ref>
<ref id="B15"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dehaene-Lambertz</surname> <given-names>G.</given-names></name></person-group> (<year>1997</year>). <article-title>Electrophysiological correlates of categorical phoneme perception in adults</article-title>. <source>NeuroReport</source> <volume>8</volume>, <fpage>919</fpage>&#x02013;<lpage>924</lpage>. <pub-id pub-id-type="doi">10.1097/00001756-199703030-00021</pub-id><pub-id pub-id-type="pmid">9141065</pub-id></citation></ref>
<ref id="B16"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Emmendorfer</surname> <given-names>A. K.</given-names></name> <name><surname>Correia</surname> <given-names>J. M.</given-names></name> <name><surname>Jansma</surname> <given-names>B. M.</given-names></name> <name><surname>Kotz</surname> <given-names>S. A.</given-names></name> <name><surname>Bonte</surname> <given-names>M.</given-names></name></person-group> (<year>2020</year>). <article-title>ERP mismatch response to phonological and temporal regularities in speech</article-title>. <source>Sci. Rep.</source> <volume>10</volume>:<fpage>9917</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-020-66824-x</pub-id><pub-id pub-id-type="pmid">32555256</pub-id></citation></ref>
<ref id="B17"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Endrass</surname> <given-names>T.</given-names></name> <name><surname>Mohr</surname> <given-names>B.</given-names></name> <name><surname>Pulvermuller</surname> <given-names>F.</given-names></name></person-group> (<year>2004</year>). <article-title>Enhanced mismatch negativity brain response after binaural word presentation</article-title>. <source>Eur. J. Neurosci.</source> <volume>19</volume>, <fpage>1653</fpage>&#x02013;<lpage>1660</lpage>. <pub-id pub-id-type="doi">10.1111/j.1460-9568.2004.03247.x</pub-id><pub-id pub-id-type="pmid">15066161</pub-id></citation></ref>
<ref id="B18"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Eriksson</surname> <given-names>A.</given-names></name> <name><surname>Traunm&#x000FC;ller</surname> <given-names>H.</given-names></name></person-group> (<year>1999</year>). <article-title>Percetion of vocal effort and speaker distance on the basis of vowel utterances</article-title>. <source>Percept. Psychophys.</source> <volume>64</volume>, <fpage>2469</fpage>&#x02013;<lpage>2472</lpage>. <pub-id pub-id-type="doi">10.3758/bf03194562</pub-id><pub-id pub-id-type="pmid">11916296</pub-id></citation></ref>
<ref id="B19"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Eriksson</surname> <given-names>A.</given-names></name> <name><surname>Traunm&#x000FC;ller</surname> <given-names>H.</given-names></name></person-group> (<year>2002</year>). <article-title>Perception of vocal effort and distance from the speaker on the basis of vowel utterances</article-title>. <source>Percept. Psychophys.</source> <volume>64</volume>, <fpage>131</fpage>&#x02013;<lpage>139</lpage>. <pub-id pub-id-type="doi">10.3758/bf03194562</pub-id><pub-id pub-id-type="pmid">11916296</pub-id></citation></ref>
<ref id="B20"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Eulitz</surname> <given-names>C.</given-names></name> <name><surname>Lahiri</surname> <given-names>A.</given-names></name></person-group> (<year>2004</year>). <article-title>Neurobiological evidence for abstract phonological representations in the mental lexicon during speech recognition</article-title>. <source>J. Cogn. Neurosci.</source> <volume>16</volume>, <fpage>577</fpage>&#x02013;<lpage>583</lpage>. <pub-id pub-id-type="doi">10.1162/089892904323057308</pub-id><pub-id pub-id-type="pmid">15185677</pub-id></citation></ref>
<ref id="B21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fletcher</surname> <given-names>H.</given-names></name> <name><surname>Munson</surname> <given-names>W. A.</given-names></name></person-group> (<year>1933</year>). <article-title>Loudness, its definition, measurement and calculation</article-title>. <source>J. Acous. Soc. Am.</source> <volume>5</volume>, <fpage>82</fpage>&#x02013;<lpage>108</lpage>. <pub-id pub-id-type="doi">10.1002/j.1538-7305.1933.tb00403.x</pub-id></citation></ref>
<ref id="B22"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Friedrich</surname> <given-names>C. K.</given-names></name> <name><surname>Lahiri</surname> <given-names>A.</given-names></name> <name><surname>Eulitz</surname> <given-names>C.</given-names></name></person-group> (<year>2008</year>). <article-title>Neurophysiological evidence for underspecified lexical representations: asymmetries with word initial variations</article-title>. <source>J. Exp. Psychol. Hum. Percept. Perform.</source> <volume>34</volume>, <fpage>1545</fpage>&#x02013;<lpage>1559</lpage>. <pub-id pub-id-type="doi">10.1037/a0012481</pub-id><pub-id pub-id-type="pmid">19045992</pub-id></citation></ref>
<ref id="B23"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Glave</surname> <given-names>R. D.</given-names></name> <name><surname>Rietveld</surname> <given-names>A. C. M.</given-names></name></person-group> (<year>1975</year>). <article-title>Is the effort dependence of speech loudness explicable on the basis of acoustical cues?</article-title> <source>J. Acoust. Soc. Am.</source> <volume>58</volume>, <fpage>875</fpage>&#x02013;<lpage>879</lpage>. <pub-id pub-id-type="doi">10.1121/1.380737</pub-id><pub-id pub-id-type="pmid">1194546</pub-id></citation></ref>
<ref id="B24"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Glave</surname> <given-names>R. D.</given-names></name> <name><surname>Rietveld</surname> <given-names>A. C. M.</given-names></name></person-group> (<year>1979</year>). <article-title>Bimodal cues for speech loudness</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>66</volume>, <fpage>1018</fpage>&#x02013;<lpage>1022</lpage>. <pub-id pub-id-type="doi">10.1121/1.383320</pub-id><pub-id pub-id-type="pmid">512212</pub-id></citation></ref>
<ref id="B25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hamamura</surname> <given-names>M.</given-names></name> <name><surname>Iwamiya</surname> <given-names>S.</given-names></name></person-group> (<year>2016</year>). <article-title>Relationship of the differences in perceived loudness of sound and the optimum listening level between males and females</article-title>. <source>Acoust. Sci. Tech.</source> <volume>37</volume>, <fpage>40</fpage>&#x02013;<lpage>41</lpage>. <pub-id pub-id-type="doi">10.1250/ast.37.40</pub-id></citation></ref>
<ref id="B26"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Henry</surname> <given-names>M. J.</given-names></name> <name><surname>Obleser</surname> <given-names>J.</given-names></name></person-group> (<year>2012</year>). <article-title>Frequency modulation entrains slow neural oscillations and optimizes human listening behavior</article-title>. <source>Proc. Natl. Acad. Sci. U S A</source> <volume>109</volume>, <fpage>20095</fpage>&#x02013;<lpage>20100</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.1213390109</pub-id><pub-id pub-id-type="pmid">23151506</pub-id></citation></ref>
<ref id="B27"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hestvik</surname> <given-names>A.</given-names></name> <name><surname>Durvasula</surname> <given-names>K.</given-names></name></person-group> (<year>2016</year>). <article-title>Neurobiological evidence for voicing underspecification in English</article-title>. <source>Brain Lang.</source> <volume>152</volume>, <fpage>28</fpage>&#x02013;<lpage>43</lpage>. <pub-id pub-id-type="doi">10.1016/j.bandl.2015.10.007</pub-id><pub-id pub-id-type="pmid">26705957</pub-id></citation></ref>
<ref id="B28"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hestvik</surname> <given-names>A.</given-names></name> <name><surname>Shinohara</surname> <given-names>Y.</given-names></name> <name><surname>Durvasula</surname> <given-names>K.</given-names></name> <name><surname>Verdonschot</surname> <given-names>R. G.</given-names></name> <name><surname>Sakai</surname> <given-names>H.</given-names></name></person-group> (<year>2020</year>). <article-title>Abstractness of human speech sound representations</article-title>. <source>Brain Res.</source> <volume>1732</volume>:<fpage>146664</fpage>. <pub-id pub-id-type="doi">10.1016/j.brainres.2020.146664</pub-id><pub-id pub-id-type="pmid">31930995</pub-id></citation></ref>
<ref id="B29"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>H&#x000F8;jlund</surname> <given-names>A.</given-names></name> <name><surname>Gebauer</surname> <given-names>L.</given-names></name> <name><surname>McGregor</surname> <given-names>W. B.</given-names></name> <name><surname>Wallentin</surname> <given-names>M.</given-names></name></person-group> (<year>2019</year>). <article-title>Context and perceptual asymmetry effects on the mismatch negativity (MMNm) to speech sounds: an MEG study</article-title>. <source>Lang. Cogn. Neurosci.</source> <volume>34</volume>, <fpage>545</fpage>&#x02013;<lpage>560</lpage>. <pub-id pub-id-type="doi">10.1080/23273798.2019.1572204</pub-id></citation></ref>
<ref id="B30"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hsu</surname> <given-names>C.-H.</given-names></name> <name><surname>Evans</surname> <given-names>J. P.</given-names></name> <name><surname>Lee</surname> <given-names>C.-Y.</given-names></name></person-group> (<year>2015</year>). <article-title>Brain responses to spoken F 0 changes: is H special?</article-title> <source>J. Phon.</source> <volume>51</volume>, <fpage>82</fpage>&#x02013;<lpage>92</lpage>. <pub-id pub-id-type="doi">10.1016/j.wocn.2015.02.003</pub-id></citation></ref>
<ref id="B32"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jacobsen</surname> <given-names>T.</given-names></name> <name><surname>Horenkamp</surname> <given-names>T.</given-names></name> <name><surname>Schr&#x000F6;ger</surname> <given-names>E.</given-names></name></person-group> (<year>2003</year>). <article-title>Preattentive memory-based comparison of sound intensity</article-title>. <source>Audiol. Neurootol.</source> <volume>8</volume>, <fpage>338</fpage>&#x02013;<lpage>346</lpage>. <pub-id pub-id-type="doi">10.1159/000073518</pub-id><pub-id pub-id-type="pmid">14566104</pub-id></citation></ref>
<ref id="B31"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jacobsen</surname> <given-names>T.</given-names></name> <name><surname>Horv&#x000E1;th</surname> <given-names>J.</given-names></name> <name><surname>Schr&#x000F6;ger</surname> <given-names>E.</given-names></name> <name><surname>Lattner</surname> <given-names>S.</given-names></name> <name><surname>Widmann</surname> <given-names>A.</given-names></name> <name><surname>Winkler</surname> <given-names>I.</given-names></name> <etal/></person-group> (<year>2004</year>). <article-title>Pre-attentive auditory processing of lexicality</article-title>. <source>Brain Lang.</source> <volume>88</volume>, <fpage>54</fpage>&#x02013;<lpage>67</lpage>. <pub-id pub-id-type="doi">10.1016/s0093-934x(03)00156-1</pub-id><pub-id pub-id-type="pmid">14698731</pub-id></citation></ref>
<ref id="B33"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Johnson</surname> <given-names>K.</given-names></name></person-group> (<year>2015</year>). <article-title>Vowel perception asymmetry in auditory and phonemic listening</article-title>. <source>UC Berkeley PhonLab Annual Report,</source> <volume>11</volume>:<fpage>26</fpage>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://escholarship.org/uc/item/21t337gh">https://escholarship.org/uc/item/21t337gh</ext-link>.</citation></ref>
<ref id="B34"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kirmse</surname> <given-names>U.</given-names></name> <name><surname>Ylinen</surname> <given-names>S.</given-names></name> <name><surname>Tervaniemi</surname> <given-names>M.</given-names></name> <name><surname>Vainio</surname> <given-names>M.</given-names></name> <name><surname>Schroger</surname> <given-names>E.</given-names></name> <name><surname>Jacobsen</surname> <given-names>T.</given-names></name></person-group> (<year>2007</year>). <article-title>Modulation of the mismatch negativity (MMN) to vowel duration changes in native speakers of Finnish and German as a result of language experience</article-title>. <source>Int. J. Psychophysiol.</source> <volume>67</volume>, <fpage>131</fpage>&#x02013;<lpage>143</lpage>. <pub-id pub-id-type="doi">10.1016/j.ijpsycho.2007.10.012</pub-id><pub-id pub-id-type="pmid">18160160</pub-id></citation></ref>
<ref id="B35"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kriengwatana</surname> <given-names>B. P.</given-names></name> <name><surname>Escudero</surname> <given-names>P.</given-names></name></person-group> (<year>2017</year>). <article-title>Directional asymmetries in vowel perception of adult nonnative listeners do not change over time with language experience</article-title>. <source>J. Speech Lang. Hear. Res.</source> <volume>60</volume>, <fpage>1088</fpage>&#x02013;<lpage>1093</lpage>. <pub-id pub-id-type="doi">10.1044/2016_JSLHR-H-16-0050</pub-id><pub-id pub-id-type="pmid">28334346</pub-id></citation></ref>
<ref id="B36"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Lahiri</surname> <given-names>A.</given-names></name> <name><surname>Reetz</surname> <given-names>H.</given-names></name></person-group> (<year>2002</year>). &#x0201C;<article-title>Underspecified recognition</article-title>,&#x0201D; in <source>Laboratory Phonology,</source> eds <person-group person-group-type="editor"><name><surname>Gussenhoven</surname> <given-names>C.</given-names></name> <name><surname>Warner</surname> <given-names>N.</given-names></name></person-group> (<publisher-name>Berlin: Mouton de Gruyter</publisher-name>), <fpage>637</fpage>&#x02013;<lpage>676</lpage>.</citation></ref>
<ref id="B37"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lahiri</surname> <given-names>A.</given-names></name> <name><surname>Reetz</surname> <given-names>H.</given-names></name></person-group> (<year>2010</year>). <article-title>Distinctive features: phonological underspecification in representation and processing</article-title>. <source>J. Phon.</source> <volume>38</volume>, <fpage>44</fpage>&#x02013;<lpage>59</lpage>. <pub-id pub-id-type="doi">10.1016/j.wocn.2010.01.002</pub-id></citation></ref>
<ref id="B38"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Langers</surname> <given-names>D. R. M.</given-names></name> <name><surname>van Dijk</surname> <given-names>P.</given-names></name> <name><surname>Schoenmaker</surname> <given-names>E. S.</given-names></name> <name><surname>Backes</surname> <given-names>W. H.</given-names></name></person-group> (<year>2007</year>). <article-title>FMRI activation in relation to sound intensity and loudness</article-title>. <source>NeuroImage</source> <volume>35</volume>, <fpage>709</fpage>&#x02013;<lpage>718</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2006.12.013</pub-id><pub-id pub-id-type="pmid">17254802</pub-id></citation></ref>
<ref id="B39"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lawyer</surname> <given-names>L. A.</given-names></name> <name><surname>Corina</surname> <given-names>D. P.</given-names></name></person-group> (<year>2018</year>). <article-title>Putting underspecification in context: ERP evidence for sparse representations in morphophonological alternations</article-title>. <source>Lang. Cogn. Neurosci.</source> <volume>33</volume>, <fpage>50</fpage>&#x02013;<lpage>64</lpage>. <pub-id pub-id-type="doi">10.1080/23273798.2017.1359635</pub-id><pub-id pub-id-type="pmid">29963576</pub-id></citation></ref>
<ref id="B40"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Lipski</surname> <given-names>S. C.</given-names></name> <name><surname>Lahiri</surname> <given-names>A.</given-names></name> <name><surname>Eulitz</surname> <given-names>C.</given-names></name></person-group> (<year>2007</year>). <article-title>Differential height specification in front vowels for German speakers and Turkish-German bilinguals: an electroencephalographic study</article-title>,&#x0201D; in <source>Proceedings of the International Congress of Phonetic Sciences XVI</source>, <publisher-loc>Saarbr&#x000FC;cken</publisher-loc>, <fpage>809</fpage>&#x02013;<lpage>812</lpage>.</citation></ref>
<ref id="B41"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Masapollo</surname> <given-names>M.</given-names></name> <name><surname>Polka</surname> <given-names>L.</given-names></name> <name><surname>M&#x000E9;nard</surname> <given-names>L.</given-names></name></person-group> (<year>2015</year>). <article-title>Asymmetries in vowel perception: effects of formant convergence and category &#x0201C;goodness</article-title>.&#x0201D; <source>J. Acoust. Soc. Am.</source> <volume>137</volume>, <fpage>2385</fpage>&#x02013;<lpage>2385</lpage>. <pub-id pub-id-type="doi">10.1016/S0167-6393(02)00105-X</pub-id></citation></ref>
<ref id="B42"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Masapollo</surname> <given-names>M.</given-names></name> <name><surname>Polka</surname> <given-names>L.</given-names></name> <name><surname>M&#x000E9;nard</surname> <given-names>L.</given-names></name></person-group> (<year>2017a</year>). <article-title>A universal bias in adult vowel perception&#x02014;by ear or by eye</article-title>. <source>Cognition</source> <volume>166</volume>, <fpage>358</fpage>&#x02013;<lpage>370</lpage>. <pub-id pub-id-type="doi">10.1159/000512876</pub-id><pub-id pub-id-type="pmid">33503641</pub-id></citation></ref>
<ref id="B44"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Masapollo</surname> <given-names>M.</given-names></name> <name><surname>Polka</surname> <given-names>L.</given-names></name> <name><surname>Molnar</surname> <given-names>M.</given-names></name> <name><surname>M&#x000E9;nard</surname> <given-names>L.</given-names></name></person-group> (<year>2017b</year>). <article-title>Directional asymmetries reveal a universal bias in adult vowel perception</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>141</volume>, <fpage>2857</fpage>&#x02013;<lpage>2869</lpage>. <pub-id pub-id-type="doi">10.1121/1.4981006</pub-id><pub-id pub-id-type="pmid">28464636</pub-id></citation></ref>
<ref id="B43"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Masapollo</surname> <given-names>M.</given-names></name> <name><surname>Polka</surname> <given-names>L.</given-names></name> <name><surname>M&#x000E9;nard</surname> <given-names>L.</given-names></name> <name><surname>Franklin</surname> <given-names>L.</given-names></name> <name><surname>Tiede</surname> <given-names>M.</given-names></name> <name><surname>Morgan</surname> <given-names>J.</given-names></name></person-group> (<year>2018</year>). <article-title>Asymmetries in unimodal visual vowel perception: the roles of oral-facial kinematics, orientation and configuration</article-title>. <source>J. Exp. Psychol. Hum. Percept. Perform.</source> <volume>44</volume>, <fpage>1103</fpage>&#x02013;<lpage>1118</lpage>. <pub-id pub-id-type="doi">10.1037/xhp0000518</pub-id><pub-id pub-id-type="pmid">29517257</pub-id></citation></ref>
<ref id="B45"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Math&#x000F4;t</surname> <given-names>S.</given-names></name> <name><surname>Schreij</surname> <given-names>D.</given-names></name> <name><surname>Theeuwes</surname> <given-names>J.</given-names></name></person-group> (<year>2012</year>). <article-title>OpenSesame: an open-source, graphical experiment builder for the social sciences</article-title>. <source>Behav. Res. Methods</source> <volume>44</volume>, <fpage>314</fpage>&#x02013;<lpage>324</lpage>. <pub-id pub-id-type="doi">10.3758/s13428-011-0168-7</pub-id><pub-id pub-id-type="pmid">22083660</pub-id></citation></ref>
<ref id="B46"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Moore</surname> <given-names>B. C. J.</given-names></name></person-group> (<year>2003</year>). <source>An Introduction to the Psychology of Hearing</source>. <publisher-loc>New York, NY</publisher-loc> <publisher-name>Academic Press</publisher-name>.</citation></ref>
<ref id="B50"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>N&#x000E4;&#x000E4;t&#x000E4;nen</surname> <given-names>R.</given-names></name></person-group> (<year>2001</year>). <article-title>The perception of speech sounds by the human brain as reflected by the mismatch negativity (MMN) and its magnetic equivalent (MMNm)</article-title>. <source>Psychophysiology</source> <volume>38</volume>, <fpage>1</fpage>&#x02013;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1017/s0048577201000208</pub-id><pub-id pub-id-type="pmid">11321610</pub-id></citation></ref>
<ref id="B49"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>N&#x000E4;&#x000E4;t&#x000E4;nen</surname> <given-names>R.</given-names></name> <name><surname>Lehtokoski</surname> <given-names>A.</given-names></name> <name><surname>Lennes</surname> <given-names>M.</given-names></name> <name><surname>Cheour</surname> <given-names>M.</given-names></name> <name><surname>Huotilainen</surname> <given-names>M.</given-names></name> <name><surname>Iivonen</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>1997</year>). <article-title>Language-specific phoneme representations revealed by electric and magnetic brain responses</article-title>. <source>Nature</source> <volume>385</volume>, <fpage>432</fpage>&#x02013;<lpage>434</lpage>. <pub-id pub-id-type="doi">10.1038/385432a0</pub-id><pub-id pub-id-type="pmid">9009189</pub-id></citation></ref>
<ref id="B47"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>N&#x000E4;&#x000E4;t&#x000E4;nen</surname> <given-names>R.</given-names></name> <name><surname>Paavilainen</surname> <given-names>P.</given-names></name> <name><surname>Reinikainen</surname> <given-names>K.</given-names></name></person-group> (<year>1989</year>). <article-title>Do event-related potentials to infrequent decrements in duration of auditory stimuli demonstrate a memory trace in man?</article-title> <source>Neurosci. Lett.</source> <volume>107</volume>, <fpage>347</fpage>&#x02013;<lpage>352</lpage>. <pub-id pub-id-type="doi">10.1016/0304-3940(89)90844-6</pub-id><pub-id pub-id-type="pmid">2616046</pub-id></citation></ref>
<ref id="B48"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>N&#x000E4;&#x000E4;t&#x000E4;nen</surname> <given-names>R.</given-names></name> <name><surname>Paavilainen</surname> <given-names>P.</given-names></name> <name><surname>Rinne</surname> <given-names>T.</given-names></name> <name><surname>Alho</surname> <given-names>K.</given-names></name></person-group> (<year>2007</year>). <article-title>The mismatch negativity (MMN) in basic research of central auditory processing: a review</article-title>. <source>Clin. Neurophysiol.</source> <volume>118</volume>, <fpage>2544</fpage>&#x02013;<lpage>2590</lpage>. <pub-id pub-id-type="doi">10.1016/j.clinph.2007.04.026</pub-id><pub-id pub-id-type="pmid">17931964</pub-id></citation></ref>
<ref id="B51"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nam</surname> <given-names>Y.</given-names></name> <name><surname>Polka</surname> <given-names>L.</given-names></name></person-group> (<year>2016</year>). <article-title>The phonetic landscape in infant consonant perception is an uneven terrain</article-title>. <source>Cognition</source> <volume>155</volume>, <fpage>57</fpage>&#x02013;<lpage>66</lpage>. <pub-id pub-id-type="doi">10.1016/j.cognition.2016.06.005</pub-id><pub-id pub-id-type="pmid">27352133</pub-id></citation></ref>
<ref id="B52"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Obleser</surname> <given-names>J.</given-names></name> <name><surname>Lahiri</surname> <given-names>A.</given-names></name> <name><surname>Eulitz</surname> <given-names>C.</given-names></name></person-group> (<year>2004</year>). <article-title>Magnetic brain response mirrors extraction of phonological features from spoken vowels</article-title>. <source>J. Cogn. Neurosci.</source> <volume>16</volume>, <fpage>31</fpage>&#x02013;<lpage>39</lpage>. <pub-id pub-id-type="doi">10.1162/089892904322755539</pub-id><pub-id pub-id-type="pmid">15006034</pub-id></citation></ref>
<ref id="B53"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Oostenveld</surname> <given-names>R.</given-names></name> <name><surname>Fries</surname> <given-names>P.</given-names></name> <name><surname>Maris</surname> <given-names>E.</given-names></name> <name><surname>Schoffelen</surname> <given-names>J.-M.</given-names></name></person-group> (<year>2011</year>). <article-title>FieldTrip: open source software for advanced analysis of MEG, EEG and invasive electrophysiological data</article-title>. <source>Comput. Intell. Neurosci.</source> <volume>2011</volume>, <fpage>1</fpage>&#x02013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1155/2011/156869</pub-id><pub-id pub-id-type="pmid">21253357</pub-id></citation></ref>
<ref id="B54"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Paavilainen</surname> <given-names>P.</given-names></name> <name><surname>Alho</surname> <given-names>K.</given-names></name> <name><surname>Reinikainen</surname> <given-names>K.</given-names></name> <name><surname>Sams</surname> <given-names>M.</given-names></name> <name><surname>N&#x000E4;&#x000E4;t&#x000E4;nen</surname> <given-names>R.</given-names></name></person-group> (<year>1991</year>). <article-title>Right hemisphere dominance of different mismatch negativities</article-title>. <source>Electroencephalogr. Clin. Neurophysiol.</source> <volume>78</volume>, <fpage>466</fpage>&#x02013;<lpage>479</lpage>. <pub-id pub-id-type="doi">10.1016/0013-4694(91)90064-b</pub-id><pub-id pub-id-type="pmid">1712282</pub-id></citation></ref>
<ref id="B55"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pakarinen</surname> <given-names>S.</given-names></name> <name><surname>Teinonen</surname> <given-names>T.</given-names></name> <name><surname>Shestakova</surname> <given-names>A.</given-names></name> <name><surname>Kwon</surname> <given-names>M. S.</given-names></name> <name><surname>Kujala</surname> <given-names>T.</given-names></name> <name><surname>H&#x000E4;m&#x000E4;l&#x000E4;inen</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2013</year>). <article-title>Fast parametric evaluation of central speech-sound processing with mismatch negativity (MMN)</article-title>. <source>Int. J. Psychophysiol.</source> <volume>87</volume>, <fpage>103</fpage>&#x02013;<lpage>110</lpage>. <pub-id pub-id-type="doi">10.1016/j.ijpsycho.2012.11.010</pub-id><pub-id pub-id-type="pmid">23201145</pub-id></citation></ref>
<ref id="B56"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Partanen</surname> <given-names>E.</given-names></name> <name><surname>Vainio</surname> <given-names>M.</given-names></name> <name><surname>Kujala</surname> <given-names>T.</given-names></name> <name><surname>Huotilainen</surname> <given-names>M.</given-names></name></person-group> (<year>2011</year>). <article-title>Linguistic multifeature MMN paradigm for extensive recording of auditory discrimination profiles</article-title>. <source>Psychophysiology</source> <volume>48</volume>, <fpage>1372</fpage>&#x02013;<lpage>1380</lpage>. <pub-id pub-id-type="doi">10.1111/j.1469-8986.2011.01214.x</pub-id><pub-id pub-id-type="pmid">21564122</pub-id></citation></ref>
<ref id="B57"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Peter</surname> <given-names>V.</given-names></name> <name><surname>McArthur</surname> <given-names>G.</given-names></name> <name><surname>Thompson</surname> <given-names>W. F.</given-names></name></person-group> (<year>2010</year>). <article-title>Effect of deviance direction and calculation method on duration and frequency mismatch negativity (MMN)</article-title>. <source>Neurosci. Lett.</source> <volume>482</volume>, <fpage>71</fpage>&#x02013;<lpage>75</lpage>. <pub-id pub-id-type="doi">10.1016/j.neulet.2010.07.010</pub-id><pub-id pub-id-type="pmid">20630487</pub-id></citation></ref>
<ref id="B58"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pettigrew</surname> <given-names>C. M.</given-names></name> <name><surname>Murdoch</surname> <given-names>B. E.</given-names></name> <name><surname>Ponton</surname> <given-names>C. W.</given-names></name> <name><surname>Finnigan</surname> <given-names>S.</given-names></name> <name><surname>Alku</surname> <given-names>P.</given-names></name> <name><surname>Kei</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2004a</year>). <article-title>Automatic auditory processing of English words as indexed by the mismatch negativity, using a multiple deviant paradigm</article-title>. <source>Ear Hear.</source> <volume>25</volume>, <fpage>284</fpage>&#x02013;<lpage>301</lpage>. <pub-id pub-id-type="doi">10.1097/01.aud.0000130800.88987.03</pub-id><pub-id pub-id-type="pmid">15179119</pub-id></citation></ref>
<ref id="B59"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pettigrew</surname> <given-names>C. M.</given-names></name> <name><surname>Murdoch</surname> <given-names>B. M.</given-names></name> <name><surname>Kei</surname> <given-names>J.</given-names></name> <name><surname>Chenery</surname> <given-names>H. J.</given-names></name> <name><surname>Sockalingam</surname> <given-names>R.</given-names></name> <name><surname>Ponton</surname> <given-names>C. W.</given-names></name> <etal/></person-group>. (<year>2004b</year>). <article-title>Processing of English words with fine acoustic contrasts and simple tones: a mismatch negativity study</article-title>. <source>J. Am. Acad. Audiol.</source> <volume>15</volume>, <fpage>47</fpage>&#x02013;<lpage>66</lpage>. <pub-id pub-id-type="doi">10.3766/jaaa.15.1.6</pub-id><pub-id pub-id-type="pmid">15030101</pub-id></citation></ref>
<ref id="B60"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Phillips</surname> <given-names>C.</given-names></name> <name><surname>Pellathy</surname> <given-names>T.</given-names></name> <name><surname>Marantz</surname> <given-names>A.</given-names></name> <name><surname>Yellin</surname> <given-names>E.</given-names></name> <name><surname>Wexler</surname> <given-names>K.</given-names></name> <name><surname>Poeppel</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2000</year>). <article-title>Auditory cortex accesses phonological categories: an MEG mismatch study</article-title>. <source>J. Cogn. Neurosci.</source> <volume>12</volume>, <fpage>1038</fpage>&#x02013;<lpage>1055</lpage>. <pub-id pub-id-type="doi">10.1162/08989290051137567</pub-id><pub-id pub-id-type="pmid">11177423</pub-id></citation></ref>
<ref id="B61"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Politzer-Ahles</surname> <given-names>S.</given-names></name> <name><surname>Schluter</surname> <given-names>K.</given-names></name> <name><surname>Wu</surname> <given-names>K.</given-names></name> <name><surname>Almeida</surname> <given-names>D.</given-names></name></person-group> (<year>2016</year>). <article-title>Asymmetries in the perception of mandarin tones: evidence from mismatch negativity</article-title>. <source>J. Exp. Psychol. Hum. Percept. Perform.</source> <volume>42</volume>, <fpage>1547</fpage>&#x02013;<lpage>1570</lpage>. <pub-id pub-id-type="doi">10.1037/xhp0000242</pub-id><pub-id pub-id-type="pmid">27195767</pub-id></citation></ref>
<ref id="B62"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Polka</surname> <given-names>L.</given-names></name> <name><surname>Bohn</surname> <given-names>O.</given-names></name></person-group> (<year>1996</year>). <article-title>A cross-language comparison of vowel perception in English-learning and German-learning infants</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>100</volume>, <fpage>577</fpage>&#x02013;<lpage>592</lpage>. <pub-id pub-id-type="doi">10.1121/1.415884</pub-id><pub-id pub-id-type="pmid">8675849</pub-id></citation></ref>
<ref id="B63"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Polka</surname> <given-names>L.</given-names></name> <name><surname>Bohn</surname> <given-names>O.-S.</given-names></name></person-group> (<year>2003</year>). <article-title>Asymmetries in vowel perception</article-title>. <source>Speech Commun.</source> <volume>41</volume>, <fpage>221</fpage>&#x02013;<lpage>231</lpage>. <pub-id pub-id-type="doi">10.1016/S0167-6393(02)00105-X</pub-id></citation></ref>
<ref id="B64"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Polka</surname> <given-names>L.</given-names></name> <name><surname>Bohn</surname> <given-names>O.-S.</given-names></name></person-group> (<year>2011</year>). <article-title>Natural Referent Vowel (NRV) framework: an emerging view of early phonetic development</article-title>. <source>J. Phon.</source> <volume>39</volume>, <fpage>467</fpage>&#x02013;<lpage>478</lpage>. <pub-id pub-id-type="doi">10.1016/j.wocn.2010.08.007</pub-id></citation></ref>
<ref id="B65"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pons</surname> <given-names>F.</given-names></name> <name><surname>Albareda-Castellot</surname> <given-names>B.</given-names></name> <name><surname>Sebasti&#x000E1;n-Gall&#x000E9;s</surname> <given-names>N.</given-names></name></person-group> (<year>2012</year>). <article-title>The interplay between input and initial biases: asymmetries in vowel perception during the first year of life: Interplay between input and initial biases</article-title>. <source>Child Dev.</source> <volume>83</volume>, <fpage>965</fpage>&#x02013;<lpage>976</lpage>. <pub-id pub-id-type="doi">10.1111/j.1467-8624.2012.01740.x</pub-id><pub-id pub-id-type="pmid">22364434</pub-id></citation></ref>
<ref id="B66"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pulverm&#x000FC;ller</surname> <given-names>F.</given-names></name> <name><surname>Kujala</surname> <given-names>T.</given-names></name> <name><surname>Shtyrov</surname> <given-names>Y.</given-names></name> <name><surname>Simola</surname> <given-names>J.</given-names></name> <name><surname>Tiitinen</surname> <given-names>H.</given-names></name> <name><surname>Alku</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2001</year>). <article-title>Memory traces for words as revealed by the mismatch negativity</article-title>. <source>NeuroImage</source> <volume>14</volume>, <fpage>607</fpage>&#x02013;<lpage>616</lpage>. <pub-id pub-id-type="doi">10.1006/nimg.2001.0864</pub-id><pub-id pub-id-type="pmid">11506534</pub-id></citation></ref>
<ref id="B67"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pulverm&#x000FC;ller</surname> <given-names>F.</given-names></name> <name><surname>Shtyrov</surname> <given-names>Y.</given-names></name> <name><surname>Ilmoniemi</surname> <given-names>R. J.</given-names></name> <name><surname>Marslen-Wilson</surname> <given-names>W. D.</given-names></name></person-group> (<year>2006</year>). <article-title>Tracking speech comprehension in space and time</article-title>. <source>NeuroImage</source> <volume>31</volume>, <fpage>1297</fpage>&#x02013;<lpage>1305</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2006.01.030</pub-id><pub-id pub-id-type="pmid">16556504</pub-id></citation></ref>
<ref id="B68"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pulverm&#x000FC;ller</surname> <given-names>F.</given-names></name> <name><surname>Shtyrov</surname> <given-names>Y.</given-names></name> <name><surname>Kujala</surname> <given-names>T.</given-names></name> <name><surname>N&#x000E4;&#x000E4;t&#x000E4;nen</surname> <given-names>R.</given-names></name></person-group> (<year>2004</year>). <article-title>Word-specific cortical activity as revealed by the mismatch negativity</article-title>. <source>Psychophysiology</source> <volume>41</volume>, <fpage>106</fpage>&#x02013;<lpage>112</lpage>. <pub-id pub-id-type="doi">10.1111/j.1469-8986.2003.00135.x</pub-id><pub-id pub-id-type="pmid">14693005</pub-id></citation></ref>
<ref id="B69"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ratcliff</surname> <given-names>R.</given-names></name></person-group> (<year>1993</year>). <article-title>Methods for dealing with reaction times outliers</article-title>. <source>Psychol. Bull.</source> <volume>114</volume>, <fpage>510</fpage>&#x02013;<lpage>532</lpage>. <pub-id pub-id-type="doi">10.1037/0033-2909.114.3.510</pub-id><pub-id pub-id-type="pmid">8272468</pub-id></citation></ref>
<ref id="B70"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sams</surname> <given-names>M.</given-names></name> <name><surname>Paavilainen</surname> <given-names>P.</given-names></name> <name><surname>Alho</surname> <given-names>K.</given-names></name></person-group> (<year>1985</year>). <article-title>Auditory freqeuncy discrimination and event-related potentials</article-title>. <source>Electroencephalogr. Clin. Neurophysiol.</source> <volume>62</volume>, <fpage>437</fpage>&#x02013;<lpage>448</lpage>. <pub-id pub-id-type="doi">10.1016/0168-5597(85)90054-1</pub-id><pub-id pub-id-type="pmid">2415340</pub-id></citation></ref>
<ref id="B71"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Savela</surname> <given-names>J.</given-names></name> <name><surname>Kujala</surname> <given-names>T.</given-names></name> <name><surname>Tuomainen</surname> <given-names>J.</given-names></name> <name><surname>Ek</surname> <given-names>M.</given-names></name> <name><surname>Aaltonen</surname> <given-names>O.</given-names></name> <name><surname>N&#x000E4;&#x000E4;t&#x000E4;nen</surname> <given-names>R.</given-names></name></person-group> (<year>2003</year>). <article-title>The mismatch negativity and reaction time as indices of the perceptual distance between the corresponding vowels of two related languages</article-title>. <source>Cogn. Brain Res.</source> <volume>16</volume>, <fpage>250</fpage>&#x02013;<lpage>256</lpage>. <pub-id pub-id-type="doi">10.1016/s0926-6410(02)00280-x</pub-id><pub-id pub-id-type="pmid">12668234</pub-id></citation></ref>
<ref id="B72"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Scharinger</surname> <given-names>M.</given-names></name></person-group> (<year>2009</year>). <article-title>Minimal representations of alternating vowels</article-title>. <source>Lingua</source> <volume>119</volume>, <fpage>1414</fpage>&#x02013;<lpage>1425</lpage>. <pub-id pub-id-type="doi">10.1177/0023830908099071</pub-id><pub-id pub-id-type="pmid">19348157</pub-id></citation></ref>
<ref id="B73"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Scharinger</surname> <given-names>M.</given-names></name> <name><surname>Bendixen</surname> <given-names>A.</given-names></name> <name><surname>Trujillo-Barreto</surname> <given-names>N. J.</given-names></name> <name><surname>Obleser</surname> <given-names>J.</given-names></name></person-group> (<year>2012a</year>). <article-title>A sparse neural code for some speech sounds but not for others</article-title>. <source>PLoS One</source> <volume>7</volume>, <fpage>e409531</fpage>&#x02013;<lpage>e4095311</lpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0040953</pub-id><pub-id pub-id-type="pmid">22815876</pub-id></citation></ref>
<ref id="B74"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Scharinger</surname> <given-names>M.</given-names></name> <name><surname>Domahs</surname> <given-names>U.</given-names></name> <name><surname>Klein</surname> <given-names>E.</given-names></name> <name><surname>Domahs</surname> <given-names>F.</given-names></name></person-group> (<year>2016</year>). <article-title>Mental representations of vowel features asymmetrically modulate activity in superior temporal sulcus</article-title>. <source>Brain Lang.</source> <volume>163</volume>, <fpage>42</fpage>&#x02013;<lpage>49</lpage>. <pub-id pub-id-type="doi">10.1016/j.bandl.2016.09.002</pub-id><pub-id pub-id-type="pmid">27669107</pub-id></citation></ref>
<ref id="B75"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Scharinger</surname> <given-names>M.</given-names></name> <name><surname>Idsardi</surname> <given-names>W. J.</given-names></name> <name><surname>Poe</surname> <given-names>S.</given-names></name></person-group> (<year>2011a</year>). <article-title>A comprehensive three-dimensional cortical map of vowel space</article-title>. <source>J. Cogn. Neurosci.</source> <volume>23</volume>, <fpage>3972</fpage>&#x02013;<lpage>3982</lpage>. <pub-id pub-id-type="doi">10.1162/jocn_a_00056</pub-id><pub-id pub-id-type="pmid">21568638</pub-id></citation></ref>
<ref id="B76"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Scharinger</surname> <given-names>M.</given-names></name> <name><surname>Lahiri</surname> <given-names>A.</given-names></name> <name><surname>Eulitz</surname> <given-names>C.</given-names></name></person-group> (<year>2010</year>). <article-title>Mismatch negativity effects of alternating vowels in morphologically complex word forms</article-title>. <source>J. Neurolinguistics</source> <volume>23</volume>, <fpage>383</fpage>&#x02013;<lpage>399</lpage>. <pub-id pub-id-type="doi">10.1016/j.jneuroling.2010.02.005</pub-id></citation></ref>
<ref id="B77"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Scharinger</surname> <given-names>M.</given-names></name> <name><surname>Merickel</surname> <given-names>J.</given-names></name> <name><surname>Riley</surname> <given-names>J.</given-names></name> <name><surname>Idsardi</surname> <given-names>W. J.</given-names></name></person-group> (<year>2011b</year>). <article-title>Neuromagnetic evidence for a featural distinction of English consonants: sensor- and source-space data</article-title>. <source>Brain Lang.</source> <volume>116</volume>, <fpage>71</fpage>&#x02013;<lpage>82</lpage>. <pub-id pub-id-type="doi">10.1016/j.bandl.2010.11.002</pub-id><pub-id pub-id-type="pmid">21185073</pub-id></citation></ref>
<ref id="B78"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Scharinger</surname> <given-names>M.</given-names></name> <name><surname>Monahan</surname> <given-names>P. J.</given-names></name> <name><surname>Idsardi</surname> <given-names>W. J.</given-names></name></person-group> (<year>2012b</year>). <article-title>Asymmetries in the processing of vowel height</article-title>. <source>J. Speech Lang. Hear. Res.</source> <volume>55</volume>, <fpage>903</fpage>&#x02013;<lpage>918</lpage>. <pub-id pub-id-type="doi">10.1044/1092-4388(2011/11-0065)</pub-id><pub-id pub-id-type="pmid">22232394</pub-id></citation></ref>
<ref id="B80"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schluter</surname> <given-names>K.</given-names></name> <name><surname>Politzer-Ahles</surname> <given-names>S.</given-names></name> <name><surname>Almeida</surname> <given-names>D.</given-names></name></person-group> (<year>2016</year>). <article-title>No place for /h/: an ERP investigation of English, fricative place features</article-title>. <source>Lang. Cogn. Neurosci.</source> <volume>31</volume>, <fpage>728</fpage>&#x02013;<lpage>740</lpage>. <pub-id pub-id-type="doi">10.1080/23273798.2016.1151058</pub-id><pub-id pub-id-type="pmid">27366758</pub-id></citation></ref>
<ref id="B79"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schluter</surname> <given-names>K. T.</given-names></name> <name><surname>Politzer-Ahles</surname> <given-names>S.</given-names></name> <name><surname>Al Kaabi</surname> <given-names>M.</given-names></name> <name><surname>Almeida</surname> <given-names>D.</given-names></name></person-group> (<year>2017</year>). <article-title>Laryngeal features are phonetically abstract: mismatch negativity evidence from Arabic, English, and Russian</article-title>. <source>Front. Psychol.</source> <volume>8</volume>:<fpage>746</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2017.00746</pub-id><pub-id pub-id-type="pmid">28555118</pub-id></citation></ref>
<ref id="B81"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schr&#x000F6;ger</surname> <given-names>E.</given-names></name></person-group> (<year>1995</year>). <article-title>Processing of auditory deviants with changes in one versus two stimulus dimensions</article-title>. <source>Psychophysiology</source> <volume>32</volume>, <fpage>55</fpage>&#x02013;<lpage>65</lpage>. <pub-id pub-id-type="doi">10.3390/ph14020079</pub-id><pub-id pub-id-type="pmid">33494166</pub-id></citation></ref>
<ref id="B82"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schr&#x000F6;ger</surname> <given-names>E.</given-names></name></person-group> (<year>1996</year>). <article-title>The influence of stimulus intensity and inter-stimulus interval on the detection of pitch and loudness changes</article-title>. <source>Electroencephalogr. Clin. Neurophysiol.</source> <volume>100</volume>, <fpage>517</fpage>&#x02013;<lpage>526</lpage>. <pub-id pub-id-type="doi">10.1016/S0168-5597(96)95576-8</pub-id><pub-id pub-id-type="pmid">8980416</pub-id></citation></ref>
<ref id="B83"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schwartz</surname> <given-names>J.-L.</given-names></name> <name><surname>Abry</surname> <given-names>C.</given-names></name> <name><surname>Bo&#x000EB;</surname> <given-names>L.-J.</given-names></name> <name><surname>M&#x000E9;nard</surname> <given-names>L.</given-names></name> <name><surname>Vall&#x000E9;e</surname> <given-names>N.</given-names></name></person-group> (<year>2005</year>). <article-title>Asymmetries in vowel perception, in the context of the dispersion-focalisation theory</article-title>. <source>Speech Commun.</source> <volume>45</volume>, <fpage>425</fpage>&#x02013;<lpage>434</lpage>. <pub-id pub-id-type="doi">10.1016/j.specom.2004.12.001</pub-id></citation></ref>
<ref id="B84"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schwartz</surname> <given-names>J.-L.</given-names></name> <name><surname>Bo&#x000EB;</surname> <given-names>L.-J.</given-names></name> <name><surname>Vall&#x000E9;e</surname> <given-names>N.</given-names></name> <name><surname>Abry</surname> <given-names>C.</given-names></name></person-group> (<year>1997</year>). <article-title>The dispersion-focalization theory of vowel systems</article-title>. <source>J. Phonetics</source> <volume>25</volume>, <fpage>255</fpage>&#x02013;<lpage>286</lpage>. <pub-id pub-id-type="doi">10.1006/jpho.1997.0043</pub-id></citation></ref>
<ref id="B85"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Sendlmeier</surname> <given-names>W. F.</given-names></name> <name><surname>Seebode</surname> <given-names>J.</given-names></name></person-group> (<year>2006</year>). <source>Formantkarten des deutschen Vokalsystems</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://www.kw.tu-berlin.de/fileadmin/a01311100/Formantkarten_des_deutschen_Vokalsystems_01.pdf">http://www.kw.tu-berlin.de/fileadmin/a01311100/Formantkarten_des_deutschen_Vokalsystems_01.pdf</ext-link>.</citation></ref>
<ref id="B86"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shtyrov</surname> <given-names>Y.</given-names></name> <name><surname>Kimppa</surname> <given-names>L.</given-names></name> <name><surname>Pulverm&#x000FC;ller</surname> <given-names>F.</given-names></name> <name><surname>Kujala</surname> <given-names>T.</given-names></name></person-group> (<year>2011</year>). <article-title>Event-related potentials reflecting the frequency of unattended spoken words: a neuronal index of connection strength in lexical memory circuits?</article-title> <source>NeuroImage</source> <volume>55</volume>, <fpage>658</fpage>&#x02013;<lpage>668</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2010.12.002</pub-id><pub-id pub-id-type="pmid">21146619</pub-id></citation></ref>
<ref id="B87"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shtyrov</surname> <given-names>Y.</given-names></name> <name><surname>Osswald</surname> <given-names>K.</given-names></name> <name><surname>Pulvermuller</surname> <given-names>F.</given-names></name></person-group> (<year>2008</year>). <article-title>Memory traces for spoken words in the brain as revealed by the hemodynamic correlate of the mismatch negativity</article-title>. <source>Cereb. Cortex</source> <volume>18</volume>, <fpage>29</fpage>&#x02013;<lpage>37</lpage>. <pub-id pub-id-type="doi">10.1093/cercor/bhm028</pub-id><pub-id pub-id-type="pmid">17412721</pub-id></citation></ref>
<ref id="B88"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shtyrov</surname> <given-names>Y.</given-names></name> <name><surname>Pulverm&#x000FC;ller</surname> <given-names>F.</given-names></name></person-group> (<year>2002</year>). <article-title>Neurophysiological evidence of memory traces for words in the human brain</article-title>. <source>Neuroreport</source> <volume>13</volume>, <fpage>521</fpage>&#x02013;<lpage>525</lpage>. <pub-id pub-id-type="doi">10.1097/00001756-200203250-00033</pub-id><pub-id pub-id-type="pmid">11930174</pub-id></citation></ref>
<ref id="B89"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Simon</surname> <given-names>E.</given-names></name> <name><surname>Sjerps</surname> <given-names>M. J.</given-names></name> <name><surname>Fikkert</surname> <given-names>P.</given-names></name></person-group> (<year>2014</year>). <article-title>Phonological representations in children&#x02019;s native and non-native lexicon</article-title>. <source>Biling. Lang. Cognit.</source> <volume>17</volume>, <fpage>3</fpage>&#x02013;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1017/S1366728912000764</pub-id></citation></ref>
<ref id="B90"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Strau&#x000DF;</surname> <given-names>A.</given-names></name> <name><surname>Kotz</surname> <given-names>S. A.</given-names></name> <name><surname>Scharinger</surname> <given-names>M.</given-names></name> <name><surname>Obleser</surname> <given-names>J.</given-names></name></person-group> (<year>2014</year>). <article-title>Alpha and theta brain oscillations index dissociable processes in spoken word recognition</article-title>. <source>NeuroImage</source> <volume>97</volume>, <fpage>387</fpage>&#x02013;<lpage>395</lpage>. <pub-id pub-id-type="doi">10.1177/1740774519877851</pub-id><pub-id pub-id-type="pmid">31647322</pub-id></citation></ref>
<ref id="B91"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Szymanski</surname> <given-names>M. D.</given-names></name> <name><surname>Yund</surname> <given-names>E. W.</given-names></name> <name><surname>Woods</surname> <given-names>D. L.</given-names></name></person-group> (<year>1999</year>). <article-title>Phonemes, intensity and attention differential effects on the mismatch negativity (MMN)</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>106</volume>, <fpage>3492</fpage>&#x02013;<lpage>3505</lpage>. <pub-id pub-id-type="doi">10.1121/1.428202</pub-id><pub-id pub-id-type="pmid">10615689</pub-id></citation></ref>
<ref id="B92"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Takegata</surname> <given-names>R.</given-names></name> <name><surname>Morotomi</surname> <given-names>T.</given-names></name></person-group> (<year>1999</year>). <article-title>Integrated neural representation of sound and temporal features in human auditory sensory memory: an event-related potential study</article-title>. <source>Neurosci. Lett.</source> <volume>274</volume>, <fpage>207</fpage>&#x02013;<lpage>210</lpage>. <pub-id pub-id-type="doi">10.1016/s0304-3940(99)00711-9</pub-id><pub-id pub-id-type="pmid">10548426</pub-id></citation></ref>
<ref id="B93"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Takegata</surname> <given-names>R.</given-names></name> <name><surname>Paavilainen</surname> <given-names>P.</given-names></name> <name><surname>N&#x000E4;&#x000E4;t&#x000E4;nen</surname> <given-names>R.</given-names></name> <name><surname>Winkler</surname> <given-names>I.</given-names></name></person-group> (<year>1999</year>). <article-title>Independent processing of changes in auditory single features and feature conjunctions in humans as indexed by the mismatch negativity</article-title>. <source>Neurosci. Lett.</source> <volume>266</volume>, <fpage>109</fpage>&#x02013;<lpage>112</lpage>. <pub-id pub-id-type="doi">10.1016/s0304-3940(99)00267-0</pub-id><pub-id pub-id-type="pmid">10353339</pub-id></citation></ref>
<ref id="B94"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Takegata</surname> <given-names>R.</given-names></name> <name><surname>Paavilainen</surname> <given-names>P.</given-names></name> <name><surname>Naatanen</surname> <given-names>R.</given-names></name> <name><surname>Winkler</surname> <given-names>I.</given-names></name></person-group> (<year>2001a</year>). <article-title>Preattentive processing of spectral, temporal and structural characteristics of acoustic regularities: a mismatch negativity study</article-title>. <source>Psychophysiology</source> <volume>38</volume>, <fpage>92</fpage>&#x02013;<lpage>98</lpage>. <pub-id pub-id-type="doi">10.1111/1469-8986.3810092</pub-id><pub-id pub-id-type="pmid">11321624</pub-id></citation></ref>
<ref id="B95"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Takegata</surname> <given-names>R.</given-names></name> <name><surname>Syssoeva</surname> <given-names>O.</given-names></name> <name><surname>Winkler</surname> <given-names>I.</given-names></name> <name><surname>Paavilainen</surname> <given-names>P.</given-names></name> <name><surname>N&#x000E4;&#x000E4;t&#x000E4;nen</surname> <given-names>R.</given-names></name></person-group> (<year>2001b</year>). <article-title>Common neural mechanism for processing onset-to-onset intervals and silent gaps in sound sequences</article-title>. <source>NeuroReport</source> <volume>12</volume>, <fpage>1783</fpage>&#x02013;<lpage>1787</lpage>. <pub-id pub-id-type="doi">10.1097/00001756-200106130-00053</pub-id><pub-id pub-id-type="pmid">11409759</pub-id></citation></ref>
<ref id="B96"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tervaniemi</surname> <given-names>M.</given-names></name> <name><surname>Schr&#x000F6;ger</surname> <given-names>E.</given-names></name> <name><surname>Saher</surname> <given-names>M.</given-names></name> <name><surname>N&#x000E4;&#x000E4;t&#x000E4;nen</surname> <given-names>R.</given-names></name></person-group> (<year>2000</year>). <article-title>Effects of spectral complexity and sound duration on automatic complex-sound pitch processing in humans&#x02014;a mismatch negativity study</article-title>. <source>Neurosci. Lett.</source> <volume>290</volume>, <fpage>66</fpage>&#x02013;<lpage>70</lpage>. <pub-id pub-id-type="doi">10.1016/s0304-3940(00)01290-8</pub-id><pub-id pub-id-type="pmid">10925176</pub-id></citation></ref>
<ref id="B97"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Todd</surname> <given-names>J.</given-names></name> <name><surname>Michie</surname> <given-names>P. T.</given-names></name></person-group> (<year>2000</year>). <article-title>Do perceived loudness cues contribute to duration mismatch negativity (MMN)?</article-title> <source>NeuroReport</source> <volume>11</volume>, <fpage>3771</fpage>&#x02013;<lpage>3774</lpage>. <pub-id pub-id-type="doi">10.1097/00001756-200011270-00035</pub-id><pub-id pub-id-type="pmid">11117488</pub-id></citation></ref>
<ref id="B98"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tuomainen</surname> <given-names>J.</given-names></name> <name><surname>Savela</surname> <given-names>J.</given-names></name> <name><surname>Obleser</surname> <given-names>J.</given-names></name> <name><surname>Aaltonen</surname> <given-names>O.</given-names></name></person-group> (<year>2013</year>). <article-title>Attention modulates the use of spectral attributes in vowel discrimination: behavioral and event-related potential evidence</article-title>. <source>Brain Res.</source> <volume>1490</volume>, <fpage>170</fpage>&#x02013;<lpage>183</lpage>. <pub-id pub-id-type="doi">10.1016/j.brainres.2012.10.067</pub-id><pub-id pub-id-type="pmid">23174416</pub-id></citation></ref>
<ref id="B99"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tyler</surname> <given-names>M. D.</given-names></name> <name><surname>Nil</surname> <given-names>M. D.</given-names></name> <name><surname>Nil</surname> <given-names>M. D.</given-names></name> <name><surname>Best</surname> <given-names>C. T.</given-names></name> <name><surname>Faber</surname> <given-names>A.</given-names></name> <name><surname>Levitt</surname> <given-names>A. G.</given-names></name></person-group> (<year>2014</year>). <article-title>Perceptual assimilation and discrimination of non-native vowel contrasts</article-title>. <source>Phonetica</source> <volume>71</volume>, <fpage>4</fpage>&#x02013;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1159/000356237</pub-id><pub-id pub-id-type="pmid">24923313</pub-id></citation></ref>
<ref id="B100"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Whelan</surname> <given-names>R.</given-names></name></person-group> (<year>2008</year>). <article-title>Effective analysis of reaction time data</article-title>. <source>Psychol. Rec.</source> <volume>58</volume>, <fpage>475</fpage>&#x02013;<lpage>482</lpage>. <pub-id pub-id-type="doi">10.1007/BF03395630</pub-id></citation></ref>
<ref id="B101"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Wiese</surname> <given-names>R.</given-names></name></person-group> (<year>2000</year>). <source>The Phonology of German</source>. <publisher-loc>Oxford</publisher-loc>: <publisher-name>Oxford University Press</publisher-name>.</citation></ref>
<ref id="B102"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wolff</surname> <given-names>C.</given-names></name> <name><surname>Schr&#x000F6;ger</surname> <given-names>E.</given-names></name></person-group> (<year>2001</year>). <article-title>Human pre-attentive auditory change-detection with single, double and triple deviations as revealed by mismatch negativity additivity</article-title>. <source>Neurosci. Lett.</source> <volume>311</volume>, <fpage>37</fpage>&#x02013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.1016/s0304-3940(01)02135-8</pub-id><pub-id pub-id-type="pmid">11585562</pub-id></citation></ref>
<ref id="B103"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yanushevskaya</surname> <given-names>I.</given-names></name> <name><surname>Gobl</surname> <given-names>C.</given-names></name> <name><surname>N&#x000ED; Chasaide</surname> <given-names>A.</given-names></name></person-group> (<year>2013</year>). <article-title>Voice quality in affect cueing: does loudness matter?</article-title> <source>Front. Psychol.</source> <volume>4</volume>:<fpage>335</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2013.00335</pub-id><pub-id pub-id-type="pmid">23785347</pub-id></citation></ref>
<ref id="B104"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yasin</surname> <given-names>I.</given-names></name></person-group> (<year>2007</year>). <article-title>Hemispheric differences in processing dichotic meaningful and non-meaningful words</article-title>. <source>Neuropsychologia</source> <volume>45</volume>, <fpage>2718</fpage>&#x02013;<lpage>2729</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuropsychologia.2007.04.009</pub-id><pub-id pub-id-type="pmid">17524432</pub-id></citation></ref>
<ref id="B105"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhao</surname> <given-names>T. C.</given-names></name> <name><surname>Masapollo</surname> <given-names>M.</given-names></name> <name><surname>Polka</surname> <given-names>L.</given-names></name> <name><surname>M&#x000E9;nard</surname> <given-names>L.</given-names></name> <name><surname>Kuhl</surname> <given-names>P. K.</given-names></name></person-group> (<year>2019</year>). <article-title>Effects of formant proximity and stimulus prototypicality on the neural discrimination of vowels: evidence from the auditory frequency-following response</article-title>. <source>Brain Lang.</source> <volume>194</volume>, <fpage>77</fpage>&#x02013;<lpage>83</lpage>. <pub-id pub-id-type="doi">10.1016/j.bandl.2019.05.002</pub-id><pub-id pub-id-type="pmid">31129300</pub-id></citation></ref>
</ref-list>
<fn-group>
<fn id="fn0001"><p><sup>1</sup><ext-link ext-link-type="uri" xlink:href="http://www.neurobs.com">http://www.neurobs.com</ext-link></p></fn>
</fn-group>
</back>
</article>
