<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Psychol.</journal-id>
<journal-title>Frontiers in Psychology</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Psychol.</abbrev-journal-title>
<issn pub-type="epub">1664-1078</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpsyg.2014.01422</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Psychology</subject>
<subj-group>
<subject>Original Research Article</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Language specific listening of Japanese geminate consonants: a cross-linguistic study</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Sadakata</surname> <given-names>Makiko</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn002"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://community.frontiersin.org/people/u/128975"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Shingai</surname> <given-names>Mizuki</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://community.frontiersin.org/people/u/196873"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Sulpizio</surname> <given-names>Simone</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<uri xlink:href="http://community.frontiersin.org/people/u/166092"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Brandmeyer</surname> <given-names>Alex</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff6"><sup>6</sup></xref>
<uri xlink:href="http://community.frontiersin.org/people/u/118082"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Sekiyama</surname> <given-names>Kaoru</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://community.frontiersin.org/people/u/115273"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Centre for Cognition, Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen</institution> <country>Nijmegen, Netherlands</country></aff>
<aff id="aff2"><sup>2</sup><institution>Institute for Logic, Language and Computation, University of Amsterdam</institution> <country>Amsterdam, Netherlands</country></aff>
<aff id="aff3"><sup>3</sup><institution>Division of Cognitive Psychology, Kumamoto University</institution> <country>Kumamoto, Japan</country></aff>
<aff id="aff4"><sup>4</sup><institution>Department of Psychology and Cognitive Science, University of Trento</institution> <country>Rovereto, Italy</country></aff>
<aff id="aff5"><sup>5</sup><institution>Fondazione ONLUS Marica De Vincenzi</institution> <country>Rovereto, Italy</country></aff>
<aff id="aff6"><sup>6</sup><institution>Max Planck Institute for Cognitive and Brain Sciences</institution> <country>Leipzig, Germany</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: <italic>Charles Clifton Jr., University of Massachusetts Amherst, USA</italic></p></fn>
<fn fn-type="edited-by"><p>Reviewed by: <italic>Hua Shu, Beijing Normal University, China; Chris Davis, University of Western Sydney, Australia</italic></p></fn>
<fn fn-type="corresp" id="fn002"><p>&#x0002A;Correspondence: <italic>Makiko Sadakata, Centre for Cognition, Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, Montessorilaan 3, 6525HE Nijmegen, Netherlands e-mail: <email>m.sadakata@donders.ru.nl</email></italic></p></fn>
<fn fn-type="other" id="fn001"><p>This article was submitted to Language Sciences, a section of the journal Frontiers in Psychology.</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>11</day>
<month>12</month>
<year>2014</year>
</pub-date>
<pub-date pub-type="collection">
<year>2014</year>
</pub-date>
<volume>5</volume>
<elocation-id>1422</elocation-id>
<history>
<date date-type="received">
<day>06</day>
<month>06</month>
<year>2014</year>
</date>
<date date-type="accepted">
<day>20</day>
<month>11</month>
<year>2014</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2014 Sadakata, Shingai, Sulpizio, Brandmeyer and Sekiyama.</copyright-statement>
<copyright-year>2014</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/"><p> This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>Various aspects of linguistic experience influence the way we segment, represent, and process speech signals. The Japanese phonetic and orthographic systems represent geminate consonants (double consonants, e.g., /ss/, /kk/) in a unique way compared to other languages: one abstract representation is used to characterize the first part of geminate consonants despite the acoustic difference between two distinct realizations of geminate consonants (silence in the case of e.g., stop consonants and elongation in the case of fricative consonants). The current study tests whether this discrepancy between abstract representations and acoustic realizations influences how native speakers of Japanese perceive geminate consonants. The experiments used pseudo words containing either the geminate consonant /ss/ or a manipulated version in which the first part was replaced by silence /_s/. The sound /_s/ is acoustically similar to /ss/, yet does not occur in everyday speech. Japanese listeners demonstrated a bias to group these two types into the same category while Italian and Dutch listeners distinguished them. The results thus confirmed that distinguishing fricative geminate consonants with silence from those with sustained frication is not crucial for Japanese native listening. Based on this observation, we propose that native speakers of Japanese tend to segment geminated consonants into two parts and that the first portion of fricative geminates is perceptually similar to a silent duration. This representation is compatible with both Japanese orthography and phonology. Unlike previous studies that were inconclusive in how native speakers segment geminate consonants, our study demonstrated a relatively strong effect of Japanese specific listening. Thus the current experimental methods may open up new lines of investigation into the relationship between development of phonological representation, orthography and speech perception.</p>
</abstract>
<kwd-group>
<kwd>phonology</kwd>
<kwd>orthography</kwd>
<kwd>perception</kwd>
<kwd>cross-linguistic</kwd>
<kwd>geminate consonants</kwd>
</kwd-group>
<counts>
<fig-count count="3"/>
<table-count count="3"/>
<equation-count count="0"/>
<ref-count count="49"/>
<page-count count="8"/>
<word-count count="0"/>
</counts>
</article-meta>
</front>
<body>
<sec>
<title>INTRODUCTION</title>
<p>Cross-linguistic studies of speech perception have unearthed language-specific differences in the perception and representation of basic speech sounds (<xref ref-type="bibr" rid="B7">Cutler, 2012</xref>). Language specific listening can be reflected by disagreements of phonemic categorical boundaries across language groups (e.g., /b/-/p/ continuum, <xref ref-type="bibr" rid="B20">Liberman et al., 1961</xref>; <xref ref-type="bibr" rid="B6">Brandmeyer et al., 2012</xref>), or by inter-language differences in the number of identified categories spanning a certain acoustic continuum (e.g., Japanese collapsing the /r/ &#x02013; /l/ continuum into one single category, <xref ref-type="bibr" rid="B5">Bradlow et al., 1997</xref>). Selective sensitivity to language-specific acoustic information also influences the perception of non-linguistic auditory materials. For example, speakers of languages that include durational contrasts in their phonemic inventory are more sensitive to changes in duration in general (<xref ref-type="bibr" rid="B32">Pickett et al., 2000</xref>; <xref ref-type="bibr" rid="B48">Ylinen et al., 2005</xref>; <xref ref-type="bibr" rid="B37">Sadakata and Sekiyama, 2011</xref>). Another well-known example is the use of language-specific segmentation/representation strategies. Here, studies using the Japanese language have been able to make considerable contributions due to its use of a distinct segmentation unit. Japanese listeners usually exhibit different segmentation strategies than, e.g., native listeners of English and/or French, especially when tested with words including sounds that could constitute a separate unit (<xref ref-type="bibr" rid="B28">Otake et al., 1993</xref>, <xref ref-type="bibr" rid="B29">1996</xref>; <xref ref-type="bibr" rid="B8">Cutler and Otake, 1994</xref>; <xref ref-type="bibr" rid="B24">Murty et al., 2007</xref>). <xref ref-type="bibr" rid="B10">Dupoux et al. (1999)</xref> and <xref ref-type="bibr" rid="B9">Dehaene-Lambertz et al. (2000)</xref> specifically highlighted the fact that Japanese phonotactics have a strong influence on speech perception; as coda consonants do not occur in Japanese (except for nasal consonants), Japanese listeners tend to perceive an illusory epenthetic vowel /u/ in a consonant cluster (e.g., /ebzo/ is often perceived as /ebuzo/ by Japanese listeners). The current study highlights another unique case in which Japanese native listeners exhibit a specific listening tendency.</p>
<p>The focus of this study is the perception of Japanese geminate consonants: double consonants such as /ss/ or /kk/. Geminate consonants are allophones of single consonants that form minimal pairs with them and which are lexically important. Geminate consonants also occur in many other languages such as Italian (<xref ref-type="bibr" rid="B32">Pickett et al., 2000</xref>; <xref ref-type="bibr" rid="B2">Bertinetto and Loporcaro, 2005</xref>; <xref ref-type="bibr" rid="B42">Tagliapietra and McQueen, 2010</xref>) and Finnish (<xref ref-type="bibr" rid="B48">Ylinen et al., 2005</xref>). The duration from the offset of a preceding vowel to the onset of a vowel immediately following the geminate consonants tends to be longer than it&#x02019;s singleton counterpart. Such consonants are often referred as long (geminate) or short (singleton) consonants (<xref ref-type="bibr" rid="B19">Kingston et al., 2009</xref>; <xref ref-type="bibr" rid="B14">Hardison and Motohashi-Saigo, 2010</xref>; <xref ref-type="bibr" rid="B42">Tagliapietra and McQueen, 2010</xref>). As such, the durational information provided by these phonemes largely accounts for the acoustic characteristics of this contrast and provides a crucial perceptual cue for native listeners (<xref ref-type="bibr" rid="B19">Kingston et al., 2009</xref>; <xref ref-type="bibr" rid="B1">Amano and Hirata, 2010</xref>) among other covariate cues, such as patterns of fundamental frequency and loudness (see for example, <xref ref-type="bibr" rid="B17">Kawahara, 2006</xref>; <xref ref-type="bibr" rid="B15">Idemaru and Guion, 2008</xref>; <xref ref-type="bibr" rid="B39">Sato et al., 2012</xref>). Moreover, learning to perceive such contrasts is not easy for speakers of non-geminate languages, suggesting that language specific listening also influences to the perception of this contrast (<xref ref-type="bibr" rid="B23">Motohashi-Saigo and Hardison, 2009</xref>; <xref ref-type="bibr" rid="B14">Hardison and Motohashi-Saigo, 2010</xref>; <xref ref-type="bibr" rid="B43">Tajima et al., 2010</xref>; <xref ref-type="bibr" rid="B36">Sadakata and McQueen, 2013</xref>). It has been previously established that Japanese native speakers show an enhanced perceptual sensitivity for consonant timing (<xref ref-type="bibr" rid="B37">Sadakata and Sekiyama, 2011</xref>). Additionally, native listeners may utilize other language-specific features of Japanese when perceiving geminates consonants, such as abstract categorical representations. The motivation for this hypothesis comes from the fact that the representation of geminate consonants in Japanese phonology and orthography deviates considerably from other languages.</p>
<p>As mentioned above, it has been convincingly demonstrated that speech segmentation is greatly influenced by phonology (<xref ref-type="bibr" rid="B28">Otake et al., 1993</xref>, <xref ref-type="bibr" rid="B29">1996</xref>; <xref ref-type="bibr" rid="B8">Cutler and Otake, 1994</xref>; <xref ref-type="bibr" rid="B10">Dupoux et al., 1999</xref>; <xref ref-type="bibr" rid="B9">Dehaene-Lambertz et al., 2000</xref>; <xref ref-type="bibr" rid="B24">Murty et al., 2007</xref>) and orthography (<xref ref-type="bibr" rid="B40">Seidenberg and Tanenhaus, 1979</xref>; <xref ref-type="bibr" rid="B49">Ziegler and Ferrand, 1998</xref>; <xref ref-type="bibr" rid="B47">Ventura et al., 2004</xref>; <xref ref-type="bibr" rid="B30">Pattamadilok et al., 2007</xref>; <xref ref-type="bibr" rid="B33">Rastle and Davis, 2008</xref>; <xref ref-type="bibr" rid="B31">Perre et al., 2009</xref>; <xref ref-type="bibr" rid="B34">Rastle et al., 2011</xref>). A number of computational models have been proposed to account for the influence of such top&#x02013;down lexical effects on phonemic decisions during speech processing. For example, the TRACE model (<xref ref-type="bibr" rid="B22">McClelland and Elman, 1986</xref>) explains that this effect takes place as a result of continuous communication between the processing of bottom&#x02013;up acoustic features and top&#x02013;down lexical representations. Later computational models, such as Shortlist (<xref ref-type="bibr" rid="B25">Norris, 1994</xref>), Merge (<xref ref-type="bibr" rid="B27">Norris et al., 2000</xref>) and Shortlist B (<xref ref-type="bibr" rid="B26">Norris and McQueen, 2008</xref>), explain this effect in a purely feed-forward manner: once bottom&#x02013;up features activate candidate representations at a higher level (e.g., input phonemes activate a list of candidate words), the lexical competition is restricted among these candidates. Despite their underlying differences, these computational models, together with the results of the previously mentioned experimental studies, suggest that speech processing is hierarchical in nature, and thus is subject to the influence of different types of linguistic representations.</p>
<p>Acoustically speaking, there are two types of Japanese obstruent geminate consonants, also known as &#x02018;sokuon&#x02019;: silent and fricative (<xref ref-type="bibr" rid="B18">Kawahara, in press</xref>). Silent geminates are produced by an abrupt suspension of articulator movements and/or by sustaining an oral closure or constriction for a short duration (e.g., stop geminate consonants, /kk/, /tt/). Fricative geminates contain sustained frication throughout their duration (e.g., /ss/). <bold>Figure <xref ref-type="fig" rid="F1">1</xref></bold> (/assu/ and /akku/) illustrates these examples. The duration of the closure or sustained frication tends to align with the duration of a mora (<xref ref-type="bibr" rid="B46">Vance, 1987</xref>; <xref ref-type="bibr" rid="B13">Han, 1994</xref>), a perceptual unit that Japanese native listeners use (<xref ref-type="bibr" rid="B28">Otake et al., 1993</xref>, <xref ref-type="bibr" rid="B29">1996</xref>; <xref ref-type="bibr" rid="B8">Cutler and Otake, 1994</xref>; <xref ref-type="bibr" rid="B24">Murty et al., 2007</xref>). <xref ref-type="bibr" rid="B44">Takahashi (1998)</xref> studied the frequency of occurrence of both types. He found that, among the 2,467 geminate consonants that appeared in the Iwanami Japanese dictionary (5th Edn), 76% were silent geminates and 24% were fricative geminates.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption><p><bold>Example waveforms of stimuli</bold>.</p></caption>
<graphic xlink:href="fpsyg-05-01422-g001.tif"/>
</fig>
<p>With regards to phonology, Japanese is one of the few languages that make use of moraic structure (<xref ref-type="bibr" rid="B46">Vance, 1987</xref>). The literature traditionally treats the first consonant of Japanese geminate consonants (&#x02018;sokuon&#x02019;) as one mora duration of the voiceless obstruent /Q/ (<xref ref-type="bibr" rid="B13">Han, 1994</xref>), while the second consonant together with the following vowel are treated as a standard mora. However, there has been no clear agreement on whether and how Japanese native listeners make use of /Q/ when perceiving geminate consonants. For example, <xref ref-type="bibr" rid="B45">Tamaoka and Terao (2004)</xref> proposed that special morae (such as the moraic nasal /N/ and /Q/) are not processed as a single mora unit, but as a part of the previous unit (e.g., CVN-CV rather than CV-N-CV). Based on this, they suggested that the latency of naming a word with CVNCV structure (which should be represented in two units) should be shorter than that with CVCVCV structure (three units). However, results regarding the moraic obstruent /Q/ were inconclusive: the authors interpreted this as /Q/ being segmented in both ways (CV-Q-CV and CVQ-CV). The current study addresses the same issue but with the method that taps directly to characteristics of abstract representations of Japanese geminate consonants (see below sections for detail). If /Q/ is perceived by native Japanese listeners as a single unit mora, then /Q/ should represent the initial portion of geminates which, as mentioned previously, can be acoustically realized in two distinct ways (silent or fricative). This is a very different way of representing geminate consonants from, e.g., one long consonant (<xref ref-type="bibr" rid="B19">Kingston et al., 2009</xref>; <xref ref-type="bibr" rid="B14">Hardison and Motohashi-Saigo, 2010</xref>; <xref ref-type="bibr" rid="B42">Tagliapietra and McQueen, 2010</xref>).</p>
<p>From an orthographic perspective, the two types of geminates are partially represented by the same Japanese syllabic character <inline-graphic xlink:href="fpsyg-05-01422-i001.jpg"/> (a small &#x02018;tsu&#x02019;). For example, &#x02018;shissaku&#x02019; (mistake) is written in four Japanese syllabic characters <inline-graphic xlink:href="fpsyg-05-01422-i002.jpg"/> (shi-tsu-sa-ku) and &#x02018;shikkaku&#x02019; (disqualified) is also written using the same character <inline-graphic xlink:href="fpsyg-05-01422-i003.jpg"/> (shi-<italic>tsu</italic>-ka-ku). Similar to /Q/, as such, one character <inline-graphic xlink:href="fpsyg-05-01422-i001.jpg"/> represents the first part of the geminate consonants while there are two acoustical types. In contrast, alphabetic languages, such as Italian, use double consonants to represent geminates: /ss/, e.g., Italian words &#x02018;asso&#x02019; (ace) vs. /kk/, e.g., &#x02018;rocca&#x02019; (fortress).</p>
<p>A discussion of the fundamental differences between phonological and orthographic systems among languages is beyond the scope of this paper. Instead, the current study asks how crucial it is for native listeners of Japanese to distinguish such acoustical differences when perceiving geminate consonants. Considering that neither the phonological nor the orthographic system suggest one-to-one correspondence between the representation and acoustic realizations &#x02013; one representation (/Q/ or ) corresponds with two geminate types (silent and fricative ones) &#x02013; we hypothesized that making perceptual distinctions between the two types of acoustic signals may not be essential for Japanese native listening. In other words, abstract representation of silent and fricative geminates may be similar for Japanese listeners. However, native speakers of other languages might treat these two as clearly separated perceptual categories.</p>
<p>We compared native listeners of Japanese and Dutch with regard to perception of spoken pseudo words containing either fricative geminates /ss/ or an artificial fricative geminate that consisted of a silence followed by a fricative singleton consonant /_/ + /s/ (<xref ref-type="bibr" rid="B38">Sadakata et al., 2012</xref>). This preliminary analysis revealed that these two stimuli indeed sounded similar for Japanese group, suggesting that it is not crucial for them to distinguish the fricative and silence geminate types, while these sounded less similar for Dutch native listeners. The results thus supported the hypothesis that Japanese make use of specific listening strategy: the two different types of geminates consonants were perceived in a similar way. However, because the Dutch language does not include geminate consonants in its inventory (<xref ref-type="bibr" rid="B4">Booij, 1995</xref>), the difference between Japanese and Dutch groups may simply reflect the presence or absence of geminates in the phoneme inventory of a given speaker&#x02019;s language. In other words, there is a possibility that native speakers of all geminate languages, such as Italian and Finish, may rely on a similar strategy as Japanese one. Because our hypothesis is more specific to Japanese native listening, the current study additionally presents the results of these experiments for native speakers of Italian.</p>
<p>The Italian language also includes phonemic contrasts between geminates and singleton consonants. However, unlike Japanese, it makes use of alphabetic representations (<xref ref-type="bibr" rid="B32">Pickett et al., 2000</xref>; <xref ref-type="bibr" rid="B42">Tagliapietra and McQueen, 2010</xref>), and does not use morae, but rather syllables for speech segmentation (e.g., <xref ref-type="bibr" rid="B41">Tabossi et al., 2000</xref>). As for segmentation of geminate consonants, Italians may share part of the strategy with Japanese, namely that it segments geminated consonants in two parts rather than treating it as a long consonant. There is a long-standing debate on whether Italian geminates are segmented into two parts or not (<xref ref-type="bibr" rid="B21">Loporcaro, 1996</xref>; <xref ref-type="bibr" rid="B2">Bertinetto and Loporcaro, 2005</xref>). The traditional phonological representation as presented in, e.g., <xref ref-type="bibr" rid="B21">Loporcaro (1996)</xref> suggests that the two consonants in Italian geminates are split, with each part belonging to the syllables before and after the geminate, respectively (e.g., asso &#x02018;ace&#x02019; is syllabified as /as.so/).</p>
<p>While it may sound similar to the Japanese two-part representation, this model has different consequences when segmenting /ss/ and /_s/. With this Italian two-part segmentation, /assu/ will be heard as /as/ + /su/ while /a_su/ will be heard as /a/ + /su/. In other words, Italian representations of two sounds (/ss/ and /_s/) are likely to be different. However, the first Italian representation (/as/ + /su/) cannot apply to Japanese native listening because Japanese phonotactics does not allow coda consonants (<xref ref-type="bibr" rid="B10">Dupoux et al., 1999</xref>; <xref ref-type="bibr" rid="B9">Dehaene-Lambertz et al., 2000</xref>). Our specific hypothesis for Japanese representation of geminate consonants is that, instead of forming a coda consonant, the first portion of voiceless geminate consonants may form an independent perceptual unit in Japanese representation that is similar to a silence. As a result, both of /assu/ and /a_su/ are represented as /a/ + /_/ + /su/. In summary, both Italian phonology and orthography differ from those of Japanese. We therefore expect the Italian group to respond similarly to the Dutch group. In order to make a direct comparison among the three language groups, we report on the data previously collected from Japanese and Dutch groups together with additional analyses.</p>
<p>Two perceptual experiments, as in <xref ref-type="bibr" rid="B38">Sadakata et al. (2012)</xref>, are reported here: a multi-voice categorization task and a multi-voice discrimination task. Although both make use of multiple voices and are therefore expected to reflect abstract linguistic representations that are independent from acoustic characteristics, the two tasks are expected to shed light on different levels of perception. The multi-voice categorization task presents four standard stimuli followed by one test stimulus (AAAAB), where participants are required to judge whether A and B belong to the same category or not (<xref ref-type="bibr" rid="B9">Dehaene-Lambertz et al., 2000</xref>). For example, the A sound might be /assu/ while the B sound could be either /a_su/ or /assu/. Participants were not instructed about the use of information regarding the presence/absence of silence when making responses. In other words, this task tests whether a participant tends to make use of or ignore the difference between /assu/ and /a_su/. Using this task, <xref ref-type="bibr" rid="B9">Dehaene-Lambertz et al. (2000)</xref> were able to capture the influence of phonotactics on perception. It is also efficient in controlling for task difficulty across the three participant groups. Conventional categorization tests often require participants to remember the target categories during subsequent categorization trials (e.g., 2AFC). This may introduce an undesired bias, as non-native speakers of Japanese may have more difficulty in remembering target categories throughout the test session, due to the fact that native speakers of Japanese produced the materials. The present task reduces the risk of this problem by not requiring participants to remember target categories.</p>
<p>The multi-voice discrimination test employed a modified version of the four-interval two-alternative forced choice task (4I-2AFC, <xref ref-type="bibr" rid="B12">Gerrits and Schouten, 2004</xref>). In this task, one deviant stimulus and three standard stimuli were presented each trial. The deviant stimuli occurred in either the second or the third position (ABAA or AABA). Participants were required to indicate the position of deviant stimuli. One crucial difference from the original study in <xref ref-type="bibr" rid="B12">Gerrits and Schouten (2004)</xref> was that each of the four words was presented in a different voice (four voices per trial). This was done in order to match this task as closely as possible to the multi-voice categorization test, as well as to ensure that participants&#x02019; sensitivity to the presence of silence was tested, independently of other acoustic cues (e.g., pitch heights, loudness envelopes). In other words, the only constant factor to determine the correct response in this task was an absence/presence of silence. In contrast to the multi-voice categorization test, this task thus requires that participants rely on the presence/absence of silence in the speech signal for making correct responses. In this way, it tests whether or not participants are sensitive to the presence/absence of silence embedded in speech signals.</p>
<p>We expected the results of the discrimination test to show that all three groups are sensitive to the presence/absence of a silence embedded in a speech sequence. More importantly, we predicted that the Japanese group would tend to ignore the difference between the two acoustic types of geminates in the categorization task, while the other groups would rely on the acoustic information. Such results would highlight the different abstract representations of fricative geminate consonants used by native speakers of Japanese, Dutch and Italian.</p>
</sec>
<sec id="s1" sec-type="materials|methods">
<title>MATERIALS AND METHODS</title>
<sec>
<title>PARTICIPANTS</title>
<p>Three groups of 16 native speakers of Japanese, Italian and Dutch took part in the two tests. Japanese participants were students of Kumamoto University (6 males and 10 females, average of 21.3 years old), Italian participants were students of Trento University (four males and 12 females, average of 25.4 years old), and Dutch participants were students of Radboud University of Nijmegen (three males and 13 females, average of 20.9 years old). The experiments were conducted in the respective countries of each of the three participant groups. All Japanese participants had followed English lessons (average 8.1 years). Some had also learned other languages, such as German, French, and Chinese. None of Italian and Dutch participants had intensive exposure to Japanese speech. Thirteen Italian participants had followed English lessons (average 7 years). Some of them spoke additional languages: French, German, Spanish, Catalan, and Esperanto. All Dutch participants had followed English lessons (average 6.8 years), and had also learned additional languages, such as French, German, Spanish, and Chinese. Written consent was obtained from all participants prior to participation. The study was carried out in accordance with the Declaration of Helsinki and Dutch legislation. The research was approved by the research ethics committee of Behavioral Research, Radboud University Nijmegen, reference number 24092009.</p>
<sec>
<title>Stimuli</title>
<p><bold>Figure <xref ref-type="fig" rid="F1">1</xref></bold> presents the waveforms of the experimental stimuli. Three types of stimuli were constructed with the following parameters: V1-/_s/-V2 (e.g., /a/ = 90 ms, /_/ = 100 ms, /s/ = 160 ms, /u/ = 100 ms), V1-/ss/-V2 (/a/ = 90 ms, /ss/ = 260 ms, /u/ = 100 ms), and V1-/kk/-V2 (/a/ = 90 ms, /_/ = 220 ms, /k/ = 40 ms, /u/ = 100 ms). These durations were chosen to correspond to the natural range of production (<xref ref-type="bibr" rid="B36">Sadakata and McQueen, 2013</xref>). Affricates (e.g., /ts/) are sometimes described as a plosive (similar to a silence) followed by a frication. However, the affricate /ts/ tends to be perceived when frication noise is much shorter (e.g., <xref ref-type="bibr" rid="B35">Repp et al., 1978</xref>). Therefore our /_s/ stimuli is, acoustically speaking, clearly distinct from both fricative and affricate geminates. Six combinations of V1 and V2 were chosen, leading to a total of 18 pseudo words (<bold>Table <xref ref-type="table" rid="T1">1</xref></bold>). Stimuli were created using six female voices and one male voice. The individual voices each corresponded to a set of recordings made by native speakers of Japanese. The speakers produced /VsV/, /VssV/, and /VkkV/ targets. Using each recording as a template, synthesized stimuli were created with identical interval durations for the three types of stimuli (/ss/, /_s/, and /kk/) using Praat (<xref ref-type="bibr" rid="B3">Boersma and Weenink, 2008</xref>). The pitch accent patterns of the two vowels /a/ and /u/ were high&#x02013;low. Onsets and offsets of each element were ramped (5 ms) to avoid clipping at the concatenation points. All materials were recorded at a sample rate of 9.6 kHz and low-pass filtered at 5 kHz in order to eliminate the low-level background noise. Average sound levels were normalized to and presented at approximately 70 dB.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>List of pseudo words.</p></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left">Stimulus type</th>
<th valign="top" align="left">Pseudo words</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">V1-/_s/-V2</td>
<td valign="top" align="left">a_su, i_se, u_sa, u_se, o_sa, o_su</td>
</tr>
<tr>
<td valign="top" align="left">V1-/ss/-V2</td>
<td valign="top" align="left">assu, isse, ussa, usse, ossa, ossu</td>
</tr>
<tr>
<td valign="top" align="left">V1-/kk/-V2</td>
<td valign="top" align="left">akku, ikke, ukka, ukke, okka, okku</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec>
<title>PROCEDURE</title>
<p>Participants took part in the multi-voice categorization test first, followed by the multi-voice discrimination test.</p>
<sec>
<title>Multi-voice categorization test</title>
<p>The multi-voice categorization task was designed in the same manner as in <xref ref-type="bibr" rid="B9">Dehaene-Lambertz et al. (2000)</xref>. Participants were presented with five words in each trial (ISI = 500 ms): four standard stimuli and one test stimulus (AAAAB). The first four items (standard) were either /ss/ or /_s/, presented by four different female voices that were randomly chosen from the six available voices (see the &#x02018;Methods&#x02019; section for details about stimuli construction). The fifth item (test) was either /ss/, /_s/, or /kk/, spoken by a male voice. The use of the male voice for the test item is in accordance with the original study. We used this large difference in voice quality between the standard and the test stimuli because it helps participants to clearly distinguish between the standards and the test item. Participants were asked to judge whether the test (fifth) word was the same as the standard items. The three test stimuli conditions are referred to as <italic>Same</italic>, <italic>Different,</italic> and <italic>Filler</italic>, respectively (see <bold>Table <xref ref-type="table" rid="T2">2</xref></bold>). Each test condition was presented 120 times, resulting in 360 trials randomly divided in 10 blocks.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Experiment conditions and predictions.</p></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left">Condition</th>
<th valign="top" align="left">Standard-Test</th>
<th valign="top" align="center" colspan="3">Predicted response<hr/></th>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<th valign="top" align="left">JP</th>
<th valign="top" align="left">IT</th>
<th valign="top" align="left">NL</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Same</td>
<td valign="top" align="left">/ss/ &#x02013; /ss/</td>
<td valign="top" align="left">Same</td>
<td valign="top" align="left">Same</td>
<td valign="top" align="left">Same</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left">/_s/ &#x02013; /_s/</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">Different</td>
<td valign="top" align="left">/ss/ &#x02013; /_s/</td>
<td valign="top" align="left">Same</td>
<td valign="top" align="left">Different</td>
<td valign="top" align="left">Different</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left">/_s/ &#x02013; /ss/</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
</tr>
<tr>
<td valign="top" align="left">Filler</td>
<td valign="top" align="left">/ss/ &#x02013; /kk/</td>
<td valign="top" align="left">Different</td>
<td valign="top" align="left">Different</td>
<td valign="top" align="left">Different</td>
</tr>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="left">/_s/ &#x02013; /kk/</td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td>
<td valign="top" align="left"></td></tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec>
<title>Multi-voice discrimination test</title>
<p>The discrimination trial sequences consisted of four words: one deviant stimulus and three standard stimuli. The deviant stimuli occurred either in the second or the third position with a 50% a priori probability (ABAA or AABA). The task was to discriminate the deviant stimulus by indicating the position of its occurrence. Two conditions were created: Silent and Stop. For both conditions, /ss/ served as standards. The Silent condition included /_s/ and the Stop condition included /kk/ as a deviant. In each trial, four words were presented (Inter-stimulus interval ISI = 500 ms). Four different female voices from the six available voices were randomly chosen for each of four words within a trial. The Silent and Stop condition were each presented 45 times, resulting in 90 trials randomly divided into six blocks.</p>
</sec>
<sec>
<title>Apparatus</title>
<p>All sounds were recorded using a linear PCM recorder (Sony PCM &#x02013; D1) with a sampling rate of 9.6 kHz in a sound attenuated booth. A DELL notebook computer with an Intel CoreDuo processer (4 GB RAM) was used to perform all experiments. Sony MDR-7506 headphones and a 15.4-inch TFT screen were used to present auditory stimuli and visual instructions, respectively. The average sound pressure level (SPL) of the headphones was adjusted to around 69 dB. The application Presentation (version 14.3, Neurobehavioral Systems) was used for presenting instructions and stimuli as well as for collecting responses. Participants responded using the computer keyboard (&#x0201C;s&#x0201D; and &#x0201C;:&#x0201D; buttons).</p>
</sec>
</sec>
</sec>
<sec>
<title>RESULTS</title>
<sec>
<title>MULTI-VOICE CATEGORIZATION TEST</title>
<p>A mixed model ANOVA with Condition (Same/Different/Filler) as a within-subjects factor and native-language (JP = Japanese, NL = Dutch, IT = Italian) as between-subjects factors on error rates indicated significant main effects of Condition [<italic>F</italic>(2,90) = 47.039; <sub>GG</sub> = 0.609; <italic>p</italic> &#x0003C; 0.0001; <inline-formula><mml:math id="M1"><mml:msubsup><mml:mi mathvariant='normal' mathcolor='black'>&#x03b7;</mml:mi><mml:mi mathvariant='normal' mathcolor='black'>p</mml:mi><mml:mn mathvariant='normal' mathcolor='black'>2</mml:mn></mml:msubsup></mml:math></inline-formula> = 0.511] as well as significant interaction between Condition and Native-language [<italic>F</italic>(4,90) = 6.3; <sub>GG</sub> = 0.609; <italic>p</italic> &#x0003C; 0.01; <inline-formula><mml:math id="M2"><mml:msubsup><mml:mi mathvariant='normal' mathcolor='black'>&#x03b7;</mml:mi><mml:mi mathvariant='normal' mathcolor='black'>p</mml:mi><mml:mn mathvariant='normal' mathcolor='black'>2</mml:mn></mml:msubsup></mml:math></inline-formula> = 0.220]. There was no significant effect of Native-language [<italic>F</italic>(2,45) = 1.718; n.s.]. Further analysis indicated several significant simple effects, which are discussed below and are indicated in <bold>Figure <xref ref-type="fig" rid="F2">2</xref></bold>. Relatively low error rates for the Filler condition for all three groups indicate that all participants performed the task correctly. As expected, Japanese data indicated significantly higher error rates for the Different condition than for the Same and Filler conditions. The Different condition presented /ss/ and /_s/, and the higher error rates for this condition confirms the hypothesis that the silent (/_s/) and fricative (/ss/) stimuli sounded similar to Japanese ears: making distinctions of these two acoustic types was not crucial for them. In contrast, a different result pattern of Same, Different and Filler conditions was observed for Dutch and Italian data as compared to that for Japanese. Significantly lower error rates were observed for the Filler condition than for the other two conditions: crucially, the error rate for the Same and Different conditions did not significantly differ for them. It is also possible that relatively higher variance in the Same condition for Dutch and Italian groups resulted in non-significant difference between the Same and Different condition, potentially due to the use of Japanese speech sounds. Nevertheless, this suggests that native speakers of Dutch and Italian perceived and indicated the difference between the silent (/_s/) and fricative (/ss/) stimuli relatively well.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption><p><bold>Mean error rates of the multi-voice categorization task.</bold> Results in all 3 conditions are indicated for Japanese, Dutch and Italian listeners (&#x0002A;&#x0002A;&#x0002A;<italic>p</italic> &#x0003C; 0.001, &#x0002A;&#x0002A;<italic>p</italic> &#x0003C; 0.01, &#x0002A;<italic>p</italic> &#x0003C; 0.05, broken line <italic>p</italic> = 0.07).</p></caption>
<graphic xlink:href="fpsyg-05-01422-g002.tif"/>
</fig>
<p><bold>Figure <xref ref-type="fig" rid="F3">3</xref></bold> presents a measure of bias, which is calculated as the difference between the error rates of the Different and Same conditions (Different &#x02013; Same). Here, the error rate of the Same condition is used as an estimate of an individual&#x02019;s task accuracy: the higher the error rate in this condition, the less accurate one is in performing the task. Therefore, subtracting the error rate of the Same condition from the Different condition indicates participants&#x02019; bias toward the same response for the Different condition. A one-way ANOVA with native-language (JP = Japanese, NL = Dutch, IT = Italian) as a between-groups factor on the bias measure indicated a strong significant main effect [<italic>F</italic>(2,47) = 7.69; <italic>p</italic> &#x0003C; 0.001; <inline-formula><mml:math id="M3"><mml:msubsup><mml:mi mathvariant='normal' mathcolor='black'>&#x03b7;</mml:mi><mml:mi mathvariant='normal' mathcolor='black'>p</mml:mi><mml:mn mathvariant='normal' mathcolor='black'>2</mml:mn></mml:msubsup></mml:math></inline-formula> = 0.255). A multiple comparisons analysis indicated that the Japanese group showed significantly greater bias than both the Italian group and the Dutch group (Tukey, both <italic>p</italic> &#x0003C; 0.01). This indicates that Japanese group had a stronger tendency toward &#x0201C;same&#x0201D; responses than the other two groups when presented with /_s/ or /ss/.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption><p><bold>Bias measure, calculated as the difference between the error rates of the Different and Same conditions (Different &#x02013; Same) for Japanese, Dutch and Italian listeners (&#x0002A;&#x0002A;<italic>p</italic> &#x0003C; 0.01)</bold>.</p></caption>
<graphic xlink:href="fpsyg-05-01422-g003.tif"/>
</fig>
</sec>
<sec>
<title>MULTI-VOICE DISCRIMINATION TEST</title>
<p>Error rates for the discrimination test are summarized in <bold>Table <xref ref-type="table" rid="T3">3</xref></bold>. A two-way mixed-model ANOVA with Condition (Silence/Stop) as a within-subjects factor and Native-language (JP/NL/IT) as a between-subjects factor on the Error rate indicated a significant effect of Condition [<italic>F</italic>(1,45) = 26.41; <italic>p</italic> &#x0003C; 0.0001; <inline-formula><mml:math id="M4"><mml:msubsup><mml:mi mathvariant='normal' mathcolor='black'>&#x03b7;</mml:mi><mml:mi mathvariant='normal' mathcolor='black'>p</mml:mi><mml:mn mathvariant='normal' mathcolor='black'>2</mml:mn></mml:msubsup></mml:math></inline-formula> = 0.37] with higher error rates for the Silence than the Stop condition. Neither the main effect of Native-language nor interaction effect was significant [<italic>F</italic>(2,45) = 0.179 n.s. and <italic>F</italic>(1,45) = 0.7436 n.s., respectively]. The results indicated that the Silence stimuli /a_su/ were more difficult to discriminate from /assu/ than the Stop stimuli /akku/ for all participants. Nevertheless, the error rates for the Silence condition were lower than 10%, suggesting that participants performed the task fairly well. Crucially, there was no difference in this sensitivity among native speakers of Japanese, Dutch and Italian.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Error rates (%) of the discrimination tests.</p></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left">Condition</th>
<th valign="top" align="left">Standard &#x02013; Deviant</th>
<th valign="top" align="left">JP</th>
<th valign="top" align="left">IT</th>
<th valign="top" align="left">NL</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Silence</td>
<td valign="top" align="left">/assu/ &#x02013; /a_su/</td>
<td valign="top" align="left">9.3 (8.8)</td>
<td valign="top" align="left">7.5 (8.2)</td>
<td valign="top" align="left">9.6 (14)</td>
</tr>
<tr>
<td valign="top" align="left">Stop</td>
<td valign="top" align="left">/assu/ &#x02013; /akku/</td>
<td valign="top" align="left">0.6 (1.3)</td>
<td valign="top" align="left">2.8 (4.8)</td>
<td valign="top" align="left">2.9 (5.1)</td></tr>
</tbody></table>
<table-wrap-foot>
<attrib><italic>Numbers in parentheses indicate SD, JP: Japanese, IT: Italian, NL: Dutch</italic>.</attrib>
</table-wrap-foot>
</table-wrap>
</sec>
</sec>
<sec>
<title>DISCUSSION</title>
<p>The current study tested whether the acoustic difference between /ss/ and /_s/ is perceptually important for Japanese native listening. Acoustically speaking, /_s/ is closer to /ss/ than /kk/. Therefore it is not too surprising if /_s/ is perceived as more similar to /ss/ than to /kk/. However, Japanese native listeners tended to categorize /ss/ and /_s/ into the same group more often than Dutch native listeners did (<xref ref-type="bibr" rid="B38">Sadakata et al., 2012</xref>). The new analysis in this study (<bold>Figure <xref ref-type="fig" rid="F3">3</xref></bold>) clearly indicated that there is a stronger &#x0201C;same&#x0201D; bias of Japanese group than Dutch group. The critical questions are, then, what do these results suggest and why such a specific way of listening came to exist. We hypothesized that this same bias is specific to Japanese native listening that is related to how Japanese language represent geminate consonants. However, the difference between Japanese and Dutch groups could simply be explained by the presence/absence of geminate consonants in participants&#x02019; native language, as the Dutch language does not include geminate consonants while Japanese does. In order to disentangle this, a third group of participants &#x02013; native Italian listeners &#x02013; was included here because Italian speech includes geminate consonants, but the manner in which Italian phonology and orthography represent geminate consonants deviates from that of Japanese (<xref ref-type="bibr" rid="B32">Pickett et al., 2000</xref>; <xref ref-type="bibr" rid="B42">Tagliapietra and McQueen, 2010</xref>). Our results indicated that the Dutch and Italian groups exhibited similar patterns, namely that they could more clearly distinguish the critical geminate sounds than the native Japanese listeners. This strengthens the idea that the observed tendency is specific to Japanese native listening: the minimal pairs of fricative and silent geminate pseudowords used as stimuli in the experiment were perceptually similar for Japanese native listeners, but not for the other two groups. The multi-voice discrimination test confirmed that the Japanese group was able to distinguish silence from frication when they were required to pay attention to the difference. Taken together, we think that the acoustic realization of the initial portion of the geminate consonants is not critical for Japanese native listening. In fact, the insertion of a silence (/_s/), which is acoustically highly unlikely, tended to be accepted as a normal fricative geminate consonant substantially more often by the Japanese participants than the Italian and Dutch participants. This can be explained if we assume that the Japanese abstract representation of geminate consonants, notably the first portion of it, is somewhat similar to a silence.</p>
<p>It is interesting to note that, using the same multi-voice sequence, the two experiments tapped different stages of linguistic processing. One crucial difference between the tasks is that the multi-voice categorization task does not force listeners to use the presence/absence of silence for making judgments while the multi-voice discrimination experiment does by requiring listeners to identify stimuli that differ from the others with respect to the presence or absence of silence. Another important difference, which may have played a role, is the level of attention level required by the two tasks. <xref ref-type="bibr" rid="B11">Fujisaki and Kawashima (1971)</xref> put forward a model with regard to speech perception that incorporates two listening modes, one that relies on categorical phonemic judgments and another based on short-term auditory memory representations of acoustic information. It is well known that traditional perceptual discrimination tasks (such as AX and ABX task) reflect phonemic categorization process in responses. However, by placing two flankers before and after the AX task, <xref ref-type="bibr" rid="B12">Gerrits and Schouten (2004)</xref> succeeded in creating a new discrimination task, namely 4I2AFC, which completely eliminated the influence of phonemic categorization on discrimination judgment; listeners in the study relied exclusively on short-term auditory memory for discrimination. The current study suggested that, even though the required judgments were not purely based on acoustic attributes of the stimuli, the multi-voice 4IAFC task seemed to induce listeners to utilize short-term auditory memory representations of acoustic information. The detection of the &#x0201C;odd-one-out&#x0201D; relied more on comparison of stimuli that were presented before/after the target word in the multi-voice discrimination (4IAFC) task, while the judgments in the multi-voice categorization task seemed to reflect more on abstract representations.</p>
<p>The tendency that Japanese listeners exhibit in the current study is compatible with both Japanese phonology and orthography. In both systems, one representation is used to characterize the first part of geminate consonants despite the acoustic difference between fricatives and stop geminate consonants: the voiceless obstruent /Q/ (<xref ref-type="bibr" rid="B46">Vance, 1987</xref>; <xref ref-type="bibr" rid="B13">Han, 1994</xref>) in phonology and the special character <inline-graphic xlink:href="fpsyg-05-01422-i001.jpg"/> in orthography. Although it is impossible to address causal relationships in this study, we believe it is still worthwhile to discuss how the relation among these systems and the observed perceptual effect could be disentangled. Using the same multi-voice categorization task with EEG measurements, <xref ref-type="bibr" rid="B9">Dehaene-Lambertz et al. (2000)</xref> demonstrated that the influence of phonotactics seems to take place at early perceptual stage of processing, namely that the phonotactic constraints are already available when our system parses and segments ongoing acoustic information. In contrast to this, the orthographic influence seems to influence later stages of processing. <xref ref-type="bibr" rid="B47">Ventura et al. (2004)</xref> and <xref ref-type="bibr" rid="B30">Pattamadilok et al. (2007)</xref> found that the effect of orthographic information processing on perceptual responses disappears when the task does not require lexical access. One approach to extend the current discussion would be to measure neural correlates associated with our geminate consonant stimuli and investigate the time course of processing differences among listener groups.</p>
<p>It has been suggested that moraic segmentation strategies in Japanese listeners develop together with the acquisition of literacy (<xref ref-type="bibr" rid="B16">Inagaki et al., 2000</xref>). However, this same study reported that the acquisition of literacy did not increase a listener&#x02019;s bias toward moraic segmentation of geminate consonants: surprisingly, 6-years old children indicated bias toward syllabic segmentation. A shift toward moraic segmentation may take place for geminate consonants too, as has been shown for the moraic nasal /N/, but at a later stage. <xref ref-type="bibr" rid="B45">Tamaoka and Terao (2004)</xref> reported that voiceless moraic obstruent /Q/ is treated either as an independent unit (CV-Q-CV) or as a part of the previous mora (CVQ-CV). These studies indicate that representations of special morae, such as /N/ and /Q/ do not develop in the same manner. In particular, the nature of the geminate consonant representations utilized by native Japanese listeners during speech processing needs further investigation. The present multi-voice categorization experiment paradigm may be able to contribute to this by asking whether learning to write <inline-graphic xlink:href="fpsyg-05-01422-i001.jpg"/> supports the formation of a common abstract representation of geminate consonants. If so, then children may show different patterns of perception before and after learning to write <inline-graphic xlink:href="fpsyg-05-01422-i004.jpg"/> with the former producing results more along the lines of those observed for the Italian and Dutch groups in this experiment.</p>
<p>To summarize, the current study highlighted a new case in which Japanese native listeners exhibit a specific listening tendency with respect to perception of Japanese geminate consonants that is similar to how Japanese phonology and orthography represent geminate consonants. Unlike previous studies that were inconclusive with respect to the manner in which native speakers segment geminate consonants, our study demonstrated a relatively strong effect of Japanese-specific listening. Our experimental methods may thus offer a new path to investigate the relationship between development of orthography and speech perception.</p>
</sec>
<sec>
<title>Conflict of Interest Statement</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</body>
<back>
<ack>
<p>The authors are grateful for support provided by for useful comments from Anne Cutler, Mirjam Broersma, Christian Hoffmann, and James McQueen. This research is supported by KAKENHI, JSPS grant-in-Aid for Challenging Exploratory Research #26590178 (leader Kaoru Sekiyama) and JSPS Grant-in-Aid for Scientific Research (A) #25240023 (leader Shuji Mori).</p>
</ack>
<ref-list>
<title>REFERENCES</title>
<ref id="B1"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Amano</surname> <given-names>S.</given-names></name> <name><surname>Hirata</surname> <given-names>Y.</given-names></name></person-group> (<year>2010</year>). <article-title>Perception and production boundaries between single and geminate stops in Japanese.</article-title> <source><italic>J. Acoust. Soc. Am.</italic></source> <volume>128</volume> <fpage>2049</fpage>&#x02013;<lpage>2058</lpage>. <pub-id pub-id-type="doi">10.1121/1.3458847</pub-id></citation></ref>
<ref id="B2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bertinetto</surname> <given-names>P. M.</given-names></name> <name><surname>Loporcaro</surname> <given-names>M.</given-names></name></person-group> (<year>2005</year>). <article-title>The sound pattern of standard Italian, as compared with the varieties spoken in Florence, Milan and Rome.</article-title> <source><italic>J. Int. Phon. Assoc.</italic></source> <volume>35</volume> <fpage>131</fpage>&#x02013;<lpage>151</lpage>. <pub-id pub-id-type="doi">10.1017/S0025100305002148</pub-id></citation></ref>
<ref id="B3"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Boersma</surname> <given-names>P.</given-names></name> <name><surname>Weenink</surname> <given-names>D.</given-names></name></person-group> (<year>2008</year>). <source><italic>Praat: Doing Phonetics By Computer (Version 5.0.36) [Computer Program].</italic></source> <comment>Available at: <ext-link ext-link-type="uri" xlink:href="http://www.praat.org/">http://www.praat.org/</ext-link> [accessed May</comment> <volume>22</volume> <issue>2012</issue>].</citation></ref>
<ref id="B4"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Booij</surname> <given-names>G.</given-names></name></person-group> (<year>1995</year>). <source><italic>The Phonology of Dutch.</italic></source> <publisher-loc>Oxford</publisher-loc>: <publisher-name>Clarendon Press</publisher-name>, <fpage>13</fpage>&#x02013;<lpage>16</lpage>.</citation></ref>
<ref id="B5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bradlow</surname> <given-names>A. R.</given-names></name> <name><surname>Pisoni</surname> <given-names>D. B.</given-names></name> <name><surname>Akahane-Yamada</surname> <given-names>R.</given-names></name> <name><surname>Tohkura</surname> <given-names>Y.</given-names></name></person-group> (<year>1997</year>). <article-title>Training Japanese listeners to identify English vertical bar r vertical bar and vertical bar l vertical bar.4. Some effects of perceptual learning on speech product.</article-title> <source><italic>J. Acoust. Soc. Am.</italic></source> <volume>101</volume> <fpage>2299</fpage>&#x02013;<lpage>2310</lpage>. <pub-id pub-id-type="doi">10.1121/1.418276</pub-id></citation></ref>
<ref id="B6"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brandmeyer</surname> <given-names>A.</given-names></name> <name><surname>Desain</surname> <given-names>P. W.</given-names></name> <name><surname>McQueen</surname> <given-names>J. M.</given-names></name></person-group> (<year>2012</year>). <article-title>Effects of native language on perceptual sensitivity to phonetic cues.</article-title> <source><italic>Neuroreport</italic></source> <volume>23</volume> <fpage>653</fpage>&#x02013;<lpage>657</lpage>. <pub-id pub-id-type="doi">10.1097/WNR.0b013e32835542cd</pub-id></citation></ref>
<ref id="B7"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cutler</surname> <given-names>A.</given-names></name></person-group> (<year>2012</year>). <source><italic>Native Listening: Language Experience and the Recognition of Spoken Words.</italic></source> <publisher-loc>Cambridge</publisher-loc>: <publisher-name>The MIT Press</publisher-name>, <fpage>1</fpage>&#x02013;<lpage>30</lpage>.</citation></ref>
<ref id="B8"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cutler</surname> <given-names>A.</given-names></name> <name><surname>Otake</surname> <given-names>T.</given-names></name></person-group> (<year>1994</year>). <article-title>Mora or phoneme &#x02013; further evidence for language-specific listening.</article-title> <source><italic>J. Mem. Lang.</italic></source> <volume>33</volume> <fpage>824</fpage>&#x02013;<lpage>844</lpage>. <pub-id pub-id-type="doi">10.1006/jmla.1994.1039</pub-id></citation></ref>
<ref id="B9"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dehaene-Lambertz</surname> <given-names>G.</given-names></name> <name><surname>Dupoux</surname> <given-names>E.</given-names></name> <name><surname>Gout</surname> <given-names>A.</given-names></name></person-group> (<year>2000</year>). <article-title>Electrophysiological correlates of phonological processing: a cross-linguistic sudy.</article-title> <source><italic>J. Cogn. Nuerosci.</italic></source> <volume>12</volume> <fpage>635</fpage>&#x02013;<lpage>647</lpage>. <pub-id pub-id-type="doi">10.1162/089892900562390</pub-id></citation></ref>
<ref id="B10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dupoux</surname> <given-names>E.</given-names></name> <name><surname>Kakehi</surname> <given-names>K.</given-names></name> <name><surname>Hirose</surname> <given-names>Y.</given-names></name> <name><surname>Pallier</surname> <given-names>C.</given-names></name> <name><surname>Mehler</surname> <given-names>J.</given-names></name></person-group> (<year>1999</year>). <article-title>Epenthetic vowels in Japanese: a perceptual illusion?</article-title> <source><italic>J. Exp. Psychol. Hum. Percept. Perform.</italic></source> <volume>25</volume> <fpage>1568</fpage>&#x02013;<lpage>1578</lpage>. <pub-id pub-id-type="doi">10.1037/0096-1523.25.6.1568</pub-id></citation></ref>
<ref id="B11"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fujisaki</surname> <given-names>H.</given-names></name> <name><surname>Kawashima</surname> <given-names>T.</given-names></name></person-group> (<year>1971</year>). <article-title>A model of the mechanisms for speech perception: quantitative analyses of categorical effects in dis- crimination.</article-title> <source><italic>Ann. Rep. Eng. Res. Inst. Fac. Eng. Univ. Tokyo</italic></source> <volume>30</volume> <fpage>59</fpage>&#x02013;<lpage>68</lpage>.</citation></ref>
<ref id="B12"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gerrits</surname> <given-names>E.</given-names></name> <name><surname>Schouten</surname> <given-names>M. E. H.</given-names></name></person-group> (<year>2004</year>). <article-title>Categorical perception depends on the discrimination task.</article-title> <source><italic>Percept. Psychophys.</italic></source> <volume>66</volume> <fpage>363</fpage>&#x02013;<lpage>376</lpage>. <pub-id pub-id-type="doi">10.3758/BF03194885</pub-id></citation></ref>
<ref id="B13"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Han</surname> <given-names>M.</given-names></name></person-group> (<year>1994</year>). <article-title>Acoustic manifestations of mora timing in Japanese.</article-title> <source><italic>J. Acoust. Soc. Am.</italic></source> <volume>96</volume> <fpage>73</fpage>&#x02013;<lpage>82</lpage>. <pub-id pub-id-type="doi">10.1121/1.410376</pub-id></citation></ref>
<ref id="B14"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hardison</surname> <given-names>D. M.</given-names></name> <name><surname>Motohashi-Saigo</surname> <given-names>M.</given-names></name></person-group> (<year>2010</year>). <article-title>Development of perception of second language Japanese geminates: role of duration, sonority, and segmentation strategy.</article-title> <source><italic>Appl. Psycholinguist.</italic></source> <volume>31</volume> <fpage>81</fpage>&#x02013;<lpage>99</lpage>. <pub-id pub-id-type="doi">10.1017/S0142716409990178</pub-id></citation></ref>
<ref id="B15"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Idemaru</surname> <given-names>K.</given-names></name> <name><surname>Guion</surname> <given-names>S. G.</given-names></name></person-group> (<year>2008</year>). <article-title>Acoustic covariants of length contrast in Japnese stops.</article-title> <source><italic>J. Int. Phon. Assoc.</italic></source> <volume>38</volume> <fpage>167</fpage>&#x02013;<lpage>186</lpage>. <pub-id pub-id-type="doi">10.1017/S0025100308003459</pub-id></citation></ref>
<ref id="B16"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Inagaki</surname> <given-names>K.</given-names></name> <name><surname>Hatano</surname> <given-names>G.</given-names></name> <name><surname>Otake</surname> <given-names>T.</given-names></name></person-group> (<year>2000</year>). <article-title>The effect of kana literacy acquisition on the speech segmentation unit used by Japanese young children.</article-title> <source><italic>J. Exp. Child Psychol.</italic></source> <volume>75</volume> <fpage>70</fpage>&#x02013;<lpage>91</lpage>. <pub-id pub-id-type="doi">10.1006/jecp.1999.2523</pub-id></citation></ref>
<ref id="B17"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kawahara</surname> <given-names>S.</given-names></name></person-group> (<year>2006</year>). <article-title>A faithfulness ranking projected from a perceptibility scale: the case of [+voice] in Japanese.</article-title> <source><italic>Language (Baltim)</italic></source> <volume>82</volume> <fpage>536</fpage>&#x02013;<lpage>574</lpage>.</citation></ref>
<ref id="B18"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kawahara</surname> <given-names>S.</given-names></name></person-group> <comment>(in press)</comment>. <article-title>&#x0201C;The phonetics of sokuon, obstruent geminates,&#x0201D; in</article-title> <source><italic>The Mouton Handbook of Japanese Language and Linguistics: Phonetics and Phonology</italic></source> <role>ed.</role> <person-group person-group-type="editor"><name><surname>Kubozono</surname> <given-names>H.</given-names></name></person-group> (<publisher-loc>Berlin</publisher-loc>: <publisher-name>Mouton Gruyter</publisher-name>).</citation></ref>
<ref id="B19"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kingston</surname> <given-names>J.</given-names></name> <name><surname>Kawahara</surname> <given-names>S.</given-names></name> <name><surname>Chambless</surname> <given-names>D.</given-names></name> <name><surname>Mash</surname> <given-names>D.</given-names></name> <name><surname>Brenner-Alsop</surname> <given-names>E.</given-names></name></person-group> (<year>2009</year>). <article-title>Contextual effects on the perception of duration.</article-title> <source><italic>J. Phon.</italic></source> <volume>37</volume> <fpage>297</fpage>&#x02013;<lpage>320</lpage>. <pub-id pub-id-type="doi">10.1016/j.wocn.2009.03.007</pub-id></citation></ref>
<ref id="B20"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liberman</surname> <given-names>A. M.</given-names></name> <name><surname>Lane</surname> <given-names>H.</given-names></name> <name><surname>Harris</surname> <given-names>K. S.</given-names></name> <name><surname>Kinney</surname> <given-names>J. A.</given-names></name></person-group> (<year>1961</year>). <article-title>Discrimination of relative onset-time of components of certain speech and nonspeech patterns.</article-title> <source><italic>J. Exp. Psychol.</italic></source> <volume>61</volume> <fpage>379</fpage>&#x02013;<lpage>388</lpage>. <pub-id pub-id-type="doi">10.1037/h0049038</pub-id></citation></ref>
<ref id="B21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Loporcaro</surname> <given-names>M.</given-names></name></person-group> (<year>1996</year>). <article-title>&#x0201C;On the analysis of geminates in Standard Italian and Italian dialects,&#x0201D; in</article-title> <source><italic>Proceedings of the Bern Workshop on Natural Phonology</italic></source> <publisher-loc>New York</publisher-loc>: <publisher-name>Mouton de Gruyter</publisher-name>, <fpage>149</fpage>&#x02013;<lpage>174</lpage>.</citation></ref>
<ref id="B22"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>McClelland</surname> <given-names>J.</given-names></name> <name><surname>Elman</surname> <given-names>J.</given-names></name></person-group> (<year>1986</year>). <article-title>The TRACE model of speech perception.</article-title> <source><italic>Cogn. Psychol. J.</italic></source> <volume>8</volume> <fpage>1</fpage>&#x02013;<lpage>86</lpage>. <pub-id pub-id-type="doi">10.1016/0010-0285(86)90015-0</pub-id></citation></ref>
<ref id="B23"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Motohashi-Saigo</surname> <given-names>M.</given-names></name> <name><surname>Hardison</surname> <given-names>D. M.</given-names></name></person-group> (<year>2009</year>). <article-title>Acquisition of L2 Japanese geminates: training with waveform displays.</article-title> <source><italic>Lang. Learn. Technol.</italic></source> <volume>13</volume> <fpage>29</fpage>&#x02013;<lpage>47</lpage>.</citation></ref>
<ref id="B24"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Murty</surname> <given-names>L.</given-names></name> <name><surname>Otake</surname> <given-names>T.</given-names></name> <name><surname>Cutler</surname> <given-names>A.</given-names></name></person-group> (<year>2007</year>). <article-title>Perceptual tests of rhythmic similarity: I. Mora rhythm.</article-title> <source><italic>Lang. Speech</italic></source> <volume>50</volume> <fpage>77</fpage>&#x02013;<lpage>99</lpage>. <pub-id pub-id-type="doi">10.1177/00238309070500010401</pub-id></citation></ref>
<ref id="B25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Norris</surname> <given-names>D.</given-names></name></person-group> (<year>1994</year>). <article-title>Shortlist: a connectionist model of continuous speech recognition.</article-title> <source><italic>Cognition</italic></source> <volume>52</volume> <fpage>189</fpage>&#x02013;<lpage>234</lpage>. <pub-id pub-id-type="doi">10.1016/0010-0277(94)90043-4</pub-id></citation></ref>
<ref id="B26"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Norris</surname> <given-names>D.</given-names></name> <name><surname>McQueen</surname> <given-names>J. M.</given-names></name></person-group> (<year>2008</year>). <article-title>Shortlist B: a Bayesian model of continuous speech recognition.</article-title> <source><italic>Psychol. Rev.</italic></source> <volume>115</volume> <fpage>357</fpage>&#x02013;<lpage>395</lpage>. <pub-id pub-id-type="doi">10.1037/0033-295X.115.2.357</pub-id></citation></ref>
<ref id="B27"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Norris</surname> <given-names>D.</given-names></name> <name><surname>McQueen</surname> <given-names>J. M.</given-names></name> <name><surname>Cutler</surname> <given-names>A.</given-names></name></person-group> (<year>2000</year>). <article-title>Merging information in speech recognition: feedback is never necessary.</article-title> <source><italic>Behav. Brain Sci.</italic></source> <volume>23</volume> <fpage>299</fpage>&#x02013;<lpage>370</lpage>. <pub-id pub-id-type="doi">10.1017/S0140525X00003241</pub-id></citation></ref>
<ref id="B28"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Otake</surname> <given-names>T.</given-names></name> <name><surname>Hatano</surname> <given-names>G.</given-names></name> <name><surname>Cutler</surname> <given-names>A.</given-names></name> <name><surname>Mehler</surname> <given-names>J.</given-names></name></person-group> (<year>1993</year>). <article-title>Mora or syllable &#x02013; speech segmentation in Japanese.</article-title> <source><italic>J. Mem. Lang.</italic></source> <volume>32</volume> <fpage>258</fpage>&#x02013;<lpage>278</lpage>. <pub-id pub-id-type="doi">10.1006/jmla.1993.1014</pub-id></citation></ref>
<ref id="B29"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Otake</surname> <given-names>T.</given-names></name> <name><surname>Yoneyama</surname> <given-names>K.</given-names></name> <name><surname>Cutler</surname> <given-names>A.</given-names></name> <name><surname>van der Lugt</surname> <given-names>A.</given-names></name></person-group> (<year>1996</year>). <article-title>The representation of Japanese moraic nasals.</article-title> <source><italic>J. Acoust. Soc. Am.</italic></source> <volume>100</volume> <fpage>3831</fpage>&#x02013;<lpage>3842</lpage>. <pub-id pub-id-type="doi">10.1121/1.417239</pub-id></citation></ref>
<ref id="B30"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pattamadilok</surname> <given-names>C.</given-names></name> <name><surname>Morais</surname> <given-names>J.</given-names></name> <name><surname>Ventura</surname> <given-names>P.</given-names></name> <name><surname>Kolinsky</surname> <given-names>R.</given-names></name></person-group> (<year>2007</year>). <article-title>The locus of the orthographic consistency effect in auditory word recognition: further evidence from French.</article-title> <source><italic>Lang. Cogn. Process.</italic></source> <volume>22</volume> <fpage>1</fpage>&#x02013;<lpage>27</lpage>. <pub-id pub-id-type="doi">10.1080/01690960601049628</pub-id></citation></ref>
<ref id="B31"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Perre</surname> <given-names>L.</given-names></name> <name><surname>Pattamadilok</surname> <given-names>C.</given-names></name> <name><surname>Montant</surname> <given-names>M.</given-names></name> <name><surname>Ziegler</surname> <given-names>J. C.</given-names></name></person-group> (<year>2009</year>). <article-title>Orthographic effects in spoken language: on-line activation or phonological restructuring?</article-title> <source><italic>Brain Res.</italic></source> <volume>1275</volume> <fpage>73</fpage>&#x02013;<lpage>80</lpage>. <pub-id pub-id-type="doi">10.1016/j.brainres.2009.04.018</pub-id></citation></ref>
<ref id="B32"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pickett</surname> <given-names>E. R.</given-names></name> <name><surname>Blumstein</surname> <given-names>S. E.</given-names></name> <name><surname>Burton</surname> <given-names>M. W.</given-names></name></person-group> (<year>2000</year>). <article-title>Effects of speaking rate on the singleton/geminate consonant contrast in Italian.</article-title> <source><italic>Phonetica</italic></source> <volume>56</volume> <fpage>135</fpage>&#x02013;<lpage>157</lpage>. <pub-id pub-id-type="doi">10.1159/000028448</pub-id></citation></ref>
<ref id="B33"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rastle</surname> <given-names>K.</given-names></name> <name><surname>Davis</surname> <given-names>M. H.</given-names></name></person-group> (<year>2008</year>). <article-title>Morphological decomposition based on the analysis of orthography.</article-title> <source><italic>Lang. Cogn. Process.</italic></source> <volume>23</volume> <fpage>942</fpage>&#x02013;<lpage>971</lpage>. <pub-id pub-id-type="doi">10.1080/01690960802069730</pub-id></citation></ref>
<ref id="B34"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rastle</surname> <given-names>K.</given-names></name> <name><surname>McCormick</surname> <given-names>S. F.</given-names></name> <name><surname>Bayliss</surname> <given-names>L.</given-names></name> <name><surname>Davis</surname> <given-names>C. J.</given-names></name></person-group> (<year>2011</year>). <article-title>Orthography influences the perception and production of speech.</article-title> <source><italic>J. Exp. Psychol. Learn. Mem. Cogn.</italic></source> <volume>37</volume> <fpage>1588</fpage>&#x02013;<lpage>1594</lpage>. <pub-id pub-id-type="doi">10.1037/a0024833</pub-id></citation></ref>
<ref id="B35"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Repp</surname> <given-names>B. H.</given-names></name> <name><surname>Liberman</surname> <given-names>A. M.</given-names></name> <name><surname>Eccardt</surname> <given-names>T.</given-names></name> <name><surname>Pesetsky</surname> <given-names>D.</given-names></name></person-group> (<year>1978</year>). <article-title>Perceptual integration of acoustic cues for stop, fricative, and affricate manner.</article-title> <source><italic>J. Exp. Psychol. Hum. Percept. Perform.</italic></source> <volume>4</volume> <fpage>621</fpage>&#x02013;<lpage>637</lpage>. <pub-id pub-id-type="doi">10.1037/0096-1523.4.4.621</pub-id></citation></ref>
<ref id="B36"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sadakata</surname> <given-names>M.</given-names></name> <name><surname>McQueen</surname> <given-names>J. M.</given-names></name></person-group> (<year>2013</year>). <article-title>High stimulus variability in nonnative speech learning supports formation of abstract categories: evidence from Japanese geminates.</article-title> <source><italic>J. Acoust. Soc. Am.</italic></source> <volume>134</volume> <fpage>1324</fpage>&#x02013;<lpage>1335</lpage>. <pub-id pub-id-type="doi">10.1121/1.4812767</pub-id></citation></ref>
<ref id="B37"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sadakata</surname> <given-names>M.</given-names></name> <name><surname>Sekiyama</surname> <given-names>K.</given-names></name></person-group> (<year>2011</year>). <article-title>Enhanced perception of various linguistic features by musicians: a cross-linguistic study.</article-title> <source><italic>Acta Psychol.</italic></source> <volume>138</volume> <fpage>1</fpage>&#x02013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1016/j.actpsy.2011.03.007</pub-id></citation></ref>
<ref id="B38"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sadakata</surname> <given-names>M.</given-names></name> <name><surname>Shingai</surname> <given-names>M.</given-names></name> <name><surname>Brandmeyer</surname> <given-names>A.</given-names></name> <name><surname>Sekiyama</surname> <given-names>K.</given-names></name></person-group> (<year>2012</year>). <article-title>&#x0201C;Perception of the moraic obstruent /Q/: a cross-linguistic study,&#x0201D; in</article-title> <source><italic>Proceedings of the Thirteenth Annual Conference of the International Speech Communication Association (Interspeech 2012)</italic>,</source> <publisher-loc>Portland, OR</publisher-loc>.</citation></ref>
<ref id="B39"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sato</surname> <given-names>Y.</given-names></name> <name><surname>Kato</surname> <given-names>M.</given-names></name> <name><surname>Macuka</surname> <given-names>R.</given-names></name></person-group> (<year>2012</year>). <article-title>Development of single/geminate obstruent discirimination by japanese infants: early integration of durational and nondurational cues.</article-title> <source><italic>Dev. Psychol.</italic></source> <volume>48</volume> <fpage>18</fpage>&#x02013;<lpage>34</lpage>. <pub-id pub-id-type="doi">10.1037/a0025528</pub-id></citation></ref>
<ref id="B40"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Seidenberg</surname> <given-names>M. S.</given-names></name> <name><surname>Tanenhaus</surname> <given-names>M. K.</given-names></name></person-group> (<year>1979</year>). <article-title>Orthographic effects on rhyme monitoring.</article-title> <source><italic>J. Exp. Psychol. Hum. Learn. Mem.</italic></source> <volume>5</volume> <fpage>546</fpage>&#x02013;<lpage>554</lpage>. <pub-id pub-id-type="doi">10.1037/0278-7393.5.6.546</pub-id></citation></ref>
<ref id="B41"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tabossi</surname> <given-names>P.</given-names></name> <name><surname>Collina</surname> <given-names>S.</given-names></name> <name><surname>Mazzetti</surname> <given-names>M.</given-names></name> <name><surname>Zoppello</surname> <given-names>M.</given-names></name></person-group> (<year>2000</year>). <article-title>Syllables in the processing of spoken Itlian.</article-title> <source><italic>J. Exp. Psychol. Hum. Percept. Perform.</italic></source> <volume>26</volume> <fpage>758</fpage>&#x02013;<lpage>775</lpage>. <pub-id pub-id-type="doi">10.1037/0096-1523.26.2.758</pub-id></citation></ref>
<ref id="B42"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tagliapietra</surname> <given-names>L.</given-names></name> <name><surname>McQueen</surname> <given-names>J. M.</given-names></name></person-group> (<year>2010</year>). <article-title>What and where in speech recognition: geminates and singletons in spoken Italian.</article-title> <source><italic>J. Mem. Lang.</italic></source> <volume>63</volume> <fpage>306</fpage>&#x02013;<lpage>323</lpage>. <pub-id pub-id-type="doi">10.1016/j.jml.2010.05.001</pub-id></citation></ref>
<ref id="B43"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tajima</surname> <given-names>K.</given-names></name> <name><surname>Kato</surname> <given-names>H.</given-names></name> <name><surname>Rothwell</surname> <given-names>A.</given-names></name> <name><surname>Akahane-Yamada</surname> <given-names>R.</given-names></name> <name><surname>Munhall</surname> <given-names>K. G.</given-names></name></person-group> (<year>2010</year>). <article-title>Training English listeners to perceive phonemic length contrasts in Japanese.</article-title> <source><italic>J. Acoust. Soc. Am.</italic></source> <volume>123</volume> <fpage>397</fpage>&#x02013;<lpage>413</lpage>. <pub-id pub-id-type="doi">10.1121/1.2804942</pub-id></citation></ref>
<ref id="B44"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Takahashi</surname> <given-names>W.</given-names></name></person-group> (<year>1998</year>). <article-title>Nihongo-no sokuon /Q/ ni tsuite (Japanese moraic obstruent /Q/).</article-title> <source><italic>J. Fac. Educ. Shinshu U.</italic></source> <volume>95</volume> <fpage>59</fpage>&#x02013;<lpage>68</lpage>.</citation></ref>
<ref id="B45"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tamaoka</surname> <given-names>K.</given-names></name> <name><surname>Terao</surname> <given-names>Y.</given-names></name></person-group> (<year>2004</year>). <article-title>Mora or syllable? Which unit do Japanese use in naming visually presented stimuli?</article-title> <source><italic>Appl. Psycholinguist.</italic></source> <volume>25</volume> <fpage>1</fpage>&#x02013;<lpage>27</lpage>. <pub-id pub-id-type="doi">10.1017/S0142716404001018</pub-id></citation></ref>
<ref id="B46"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vance</surname> <given-names>T. J.</given-names></name></person-group> (<year>1987</year>). <italic>An Introduction to Japanese Phonology</italic> (<publisher-loc>Albany, NY</publisher-loc>: <publisher-name>State University of New York Press</publisher-name>), <fpage>34</fpage>&#x02013;<lpage>47</lpage>.</citation></ref>
<ref id="B47"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ventura</surname> <given-names>P.</given-names></name> <name><surname>Morais</surname> <given-names>J.</given-names></name> <name><surname>Pattamadilok</surname> <given-names>C.</given-names></name> <name><surname>Kolinsky</surname> <given-names>R.</given-names></name></person-group> (<year>2004</year>). <article-title>The locus of the orthographic consistency effect in auditory word recognition.</article-title> <source><italic>Lang. Cogn. Proc.</italic></source> <volume>19</volume> <fpage>57</fpage>&#x02013;<lpage>95</lpage>. <pub-id pub-id-type="doi">10.1080/01690960344000134</pub-id></citation></ref>
<ref id="B48"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ylinen</surname> <given-names>S.</given-names></name> <name><surname>Shestakova</surname> <given-names>A.</given-names></name> <name><surname>Alku</surname> <given-names>P.</given-names></name> <name><surname>Huotilainen</surname> <given-names>M.</given-names></name></person-group> (<year>2005</year>). <article-title>The perception of phonological quantity based on durational cues by native speakers, second-language users and nonspeakers of Finnish.</article-title> <source><italic>Lang. Speech</italic></source> <volume>48</volume> <fpage>313</fpage>&#x02013;<lpage>338</lpage>. <pub-id pub-id-type="doi">10.1177/00238309050480030401</pub-id></citation></ref>
<ref id="B49"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ziegler</surname> <given-names>J. C.</given-names></name> <name><surname>Ferrand</surname> <given-names>L.</given-names></name></person-group> (<year>1998</year>). <article-title>Orthography shapes the perception of speech: the consistency effect in auditory word recognition.</article-title> <source><italic>Psychon. Bull. Rev.</italic></source> <volume>5</volume> <fpage>683</fpage>&#x02013;<lpage>689</lpage>. <pub-id pub-id-type="doi">10.3758/BF03208845</pub-id></citation></ref>
</ref-list>
</back>
</article>