Skip to main content

ORIGINAL RESEARCH article

Front. Hum. Neurosci., 23 December 2016
Sec. Speech and Language
Volume 10 - 2016 | https://doi.org/10.3389/fnhum.2016.00652

Neural Correlates of Indicators of Sound Change in Cantonese: Evidence from Cortical and Subcortical Processes

  • 1Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Hong Kong, China
  • 2School of Psychology and Clinical Language Sciences, University of Reading, Reading, UK
  • 3The MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Penrith, NSW, Australia
  • 4Brain and Mind Institute, The Chinese University of Hong Kong, Hong Kong, China
  • 5The Chinese University of Hong Kong—Utrecht University Joint Center for Language, Mind and Brain, Hong Kong, China

Across time, languages undergo changes in phonetic, syntactic, and semantic dimensions. Social, cognitive, and cultural factors contribute to sound change, a phenomenon in which the phonetics of a language undergo changes over time. Individuals who misperceive and produce speech in a slightly divergent manner (called innovators) contribute to variability in the society, eventually leading to sound change. However, the cause of variability in these individuals is still unknown. In this study, we examined whether such misperceptions are represented in neural processes of the auditory system. We investigated behavioral, subcortical (via FFR), and cortical (via P300) manifestations of sound change processing in Cantonese, a Chinese language in which several lexical tones are merging. Across the merging categories, we observed a similar gradation of speech perception abilities in both behavior and the brain (subcortical and cortical processes). Further, we also found that behavioral evidence of tone merging correlated with subjects' encoding at the subcortical and cortical levels. These findings indicate that tone-merger categories, that are indicators of sound change in Cantonese, are represented neurophysiologically with high fidelity. Using our results, we speculate that innovators encode speech in a slightly deviant neurophysiological manner, and thus produce speech divergently that eventually spreads across the community and contributes to sound change.

Introduction

Language is a dynamic and adaptive system that undergoes diachronic and synchronic changes (Beckner et al., 2009). Some of the well-documented changes across languages include the great vowel chain shifts between the 14th and 16th century (Wolfe, 1972), phonological mergers in Sinitic languages (Shen, 1997) and lexical borrowing between languages (Bloomfield, 1933; Cheng, 1987). In addition to documenting change, research studies have also investigated the sources of change, predominantly by examining the impact of social factors (Labov, 1980, 1990, 2006; Haeri, 1996; Eckert, 2000) on language. In recent years, a new set of studies have begun to focus on speaker-internal factors that could drive change, including individual differences in perceptual and cognitive processes (Witkin et al., 1975; Messick, 1976; Ausburn and Ausburn, 1978; Labov, 2001; Yu, 2010, 2013; Browman and Goldstein, 2014). The aim of the current study is to further investigate speaker-internal factors by examining how individual differences in neurophysiological processing may be associated with language change. The sound system of Cantonese is undergoing a course of change in which some speech categories are being merged. Most interestingly, the extent of merging differs across individuals and across speech categories. This gives rise to the opportunity for the current study to connect the extent of merging to differences in neurophysiological responses across the auditory neural pathway.

For decades, a large body of influential work on language change by Labov and others has found social factors to be predominant driving forces of change. Among the factors investigated are the effects of local communities (Eckert, 2000), socio-economic classes (Trudgill, 1974; Weinberg, 1974; Labov, 1980, 2006; Haeri, 1996; Velde et al., 2001), gender (Gauchat, 1905; Hermann, 1929; Labov, 1990, 2001; Yu, 2010), regional dialect (Labov, 2010), and frequency of usage (Hooper, 1976; Pierrehumbert, 2001; Abramowicz, 2007; Zhao and Jurafsky, 2009). Apart from social factors, cognitive differences among the individuals have also been considered as important contributors toward sound change (Ohala, 1990; Labov, 2010). It has been postulated that in a society, there are innovators who are individuals with a slightly deviant style of speech perception and production, and early adopters who adopt their deviant style (Milroy and Milroy, 1985). As innovators and early adopters interact, there is a spread of divergent speech characteristics in the society that during the course of time leads to modification of norm ultimately leading to sound change. Research studies also ask what speaker-internal factors may lead certain individuals in society to become innovators and others early adopters. Perceptual and cognitive factors are likely to contribute to such biases (Yu, 2010, 2013). In particular, individual differences in speech perception can be a particularly important driving force (Yu, 2010; Ou et al., 2015). As certain individuals misperceive speech, the misperception could lead to adjustments in production norms leading to development of new variants. When these variants are associated with social significance and spread to the rest of the speech community, sound change occurs. Ohala (1993) reports that the three main mechanisms responsible for sound change are hypocorrection in which a listener under-corrects coarticulatory effects; hypercorrection in which a listener overcorrects coarticulatory effects; and confusion between the similar acoustical sounds that occurs due to failure of a listener to retrieve a feature in one sound but not in the other. One possibility is that the listeners who hyper/hypocorrect are the innovators who perceive and produce in a divergent manner and thus, modify the norms of the community.

Are these confusions or misperceptions represented at the subcortical and cortical levels? For speech to be perceived, it must undergo stages of processing in both the peripheral (outer, middle and inner ear) and central (auditory nerve to brainstem to the cortex) components of the auditory pathway. The auditory brainstem and cortex are interconnected via numerous afferent and efferent neuronal pathways. The brainstem is a complex hierarchical structure consisting of cochlear nucleus, superior olivary complex, lateral lemniscus, and inferior colliculus interconnected via afferent, efferent, ipsilateral, and/or contralateral nerve fibers (Chandrasekaran and Kraus, 2010). The brainstem acts as a hub for auditory processing, and its activity can be measured using the frequency following response (FFR). FFR has been used as a metric of auditory system plasticity due to language experience (Krishnan et al., 2005, 2009; Swaminathan et al., 2008) and musical training (Musacchia et al., 2007, 2008; Wong et al., 2007; Bidelman et al., 2009). Furthermore, the brainstem with its sharp phase-locking characteristics (Galbraith et al., 1995) and top-down feedback from the cortex via efferent pathways, acts as an active contributor rather than a passive relay station in speech sound processing (Chandrasekaran et al., 2014). It has been found that clinical populations with disorders of speech, language and reading have impaired subcortical representation of sounds (Wible et al., 2005; Hornickel et al., 2009; Russo et al., 2009, 2010; Anderson et al., 2013). Given the importance of brainstem in speech sound encoding, we used FFR as a tool to study the distinction of tone categories in the brainstem. A deficiency in brainstem encoding may lead to defective representation of sounds in the cortex and vice versa (Chandrasekaran et al., 2014).

Apart from subcortical measures, cortical measures of auditory processing have also been found to be sensitive in determining the effects of long-term language experience (Buchwald et al., 1994; Chandrasekaran et al., 2007, 2009a; Zheng et al., 2012). Chandrasekaran et al. (2007) used mismatched negativity (MMN) and found that Chinese and English speakers differed on their pre-attentive ability to discriminate Mandarin lexical tone pairs. Further, Chandrasekaran et al. (2009a) also found that Chinese speakers processed lexical tones more faithfully than musicians. They attributed these findings to their long-term language experience. Though the effects of linguistic experience can be gauged with MMN or even earlier components, they may be driven by experience-dependent acoustic features (Maiste et al., 1995; Chandrasekaran et al., 2007, 2009b) rather than perception. In contrast, P300 is known to indicate discrimination of speech sounds based on phonological information (Maiste et al., 1995; Frenck-Mestre et al., 2005). Because of different attentional requirements, MMN and P300 index different stages of processing in the brain. MMN is a result of automatic change detection regardless of whether the change occurs in the acoustic or phonemic domain (Näätänen et al., 2007) and is a relatively early-stage phenomenon occurring in the latency range of 100–300 ms. P300, in contrast, is an event related potential (ERP) usually elicited in the range of 250–500 ms in an oddball paradigm where attention needs to be allocated to the deviant stimuli. In a P300 task, memory schemas are built with the incoming standard stimuli and every time a rare or deviant stimulus appears, these schemas are updated. P300 amplitude and latency is usually modulated by arousal levels and the amount of effort required in detection of the deviant stimuli (Polich, 2007). The more effortful the task, the longer the P300 latency and the higher the amplitude (Luck, 2005). P300 has been used to study the effects of linguistic experience (Buchwald et al., 1994; Zheng et al., 2012). Buchwald et al. (1994) studied whether Japanese listeners could differentiate/r/ and /l/ phonemes that are contrasted in English but not in Japanese. They found that adult native speakers of Japanese had reduced P300 for detecting [r] and [l] sounds. Zheng et al. (2012) found that Cantonese speakers had enhanced P300 amplitudes relative to their Mandarin counterparts on a lexical tone oddball discrimination task, mainly due to long-term experience with the more complex tone space in Cantonese. As categorical perception of lexical tones can be well represented neurophysiologically using P300 (Zheng et al., 2012), we used this ERP technique in the current study to probe active language processing at the cortical level. In sum, we employed a combination of FFR and P300 to probe passive (subcortical) and active (cortical) levels of auditory processing to investigate neural representations of the phenomenon of sound change.

It is known that encoding at the auditory brainstem is shaped by language experience and tone language speakers encode lexical tones more robustly than non-tone language speakers (Krishnan et al., 2005, 2009; Swaminathan et al., 2008). Krishnan et al. (2005, 2009, 2010) argue that local mechanisms in the brainstem circuitry show plasticity toward long-term experience. On the other hand, there is also evidence that brainstem encoding is driven by top-down cortical feedback (Suga, 2008; Bajo et al., 2010). Recently, it has been found that the effects visible at the brainstem originate at the level of cortex and are driven via corticofugal pathways from cortex to the brainstem (Suga, 2008; Parbery-Clark et al., 2009; Krizman et al., 2012; Song et al., 2012; Hairston et al., 2013; Skoe and Kraus, 2013). As literature supports the effect of tone language experience both at the brainstem and cortex, it is important to probe these levels of the auditory pathway using FFR and P300 in order to understand whether sound change is represented at the brainstem and/or cortical level.

In order to investigate the neural representations of sound change in the auditory pathway, we examined the neural responses to speech categories in Cantonese that are undergoing various degrees of merging. Cantonese is a Chinese language that has six lexical tones (high-level, mid-level, low-level, high-rising, low-rising, and falling), which are pitch patterns used to distinguish word meanings (e.g., the syllable /ji/ means “chair” when spoken with high-rising tone 2 and “ear” when spoken with low-rising tone 5). What makes Cantonese particularly useful for studying the neural basis of sound change is that some of the tone categories are in the process of undergoing merging to various degrees. Merging leads to the disappearance and emergence of speech sound categories in a language. The merging phenomenon may be observed in the English of some speakers in the southern United States where they have lost distinction between “pen” and “pin” (the sound “e” has been lost) (Labov et al., 2005). In Cantonese, several degrees of merging can be observed across the tone categories. Degrees of merging are determined based on the prevalence of confusions in native Cantonese speakers' perception and/or production (Law et al., 2013; Mok et al., 2013). The two rising lexical tones of Cantonese, Tones 2 and 5 (denoted by the symbols T2 and T5, respectively) are confused most often and thus constitute a full-merger pair, followed by Tones 4 (falling pitch) and 6 (low-level pitch) forming a near-merger pair (T4/T6), and Tones 3 (mid-level pitch) and 6 (low-level pitch) form a quasi-merger pair (T3/T6) with the least confusion compared to the other merging pairs (Fung and Wong, 2011). Conversely, Tones 1 (high-level pitch) and 2 (high-rising pitch) form a non-merger pair (T1/T2) causing no confusion in perception. Further, lexical tones are especially useful for investigating subcortical auditory processing because the brainstem response is phase-locked to the fundamental frequency (F0), the acoustic correlate of lexical tone.

In the current study, we explored the neural representations of sound change by investigating tone merging at behavioral, brainstem and cortical levels. Given that perceptual and cognitive factors are thought to contribute to sound change (Yu, 2010, 2013), we hypothesized individual variability in the behavioral, brainstem and/or cortical encoding of speech sounds. Specifically, we predicted a high correlation between the behavioral and neurophysiological manifestations of merging. Additionally, we predicted a similar gradation of degree of tone merging in the behavior and brainstem/cortical representations.

To evaluate tone merging at the level of the brainstem, we recorded FFR from adult native Cantonese speakers (n = 30) using the six lexical tones of Cantonese. The magnitude of merging in the brainstem was determined using response-to-response correlation of FFR pitch contours of the tones in the relevant tone pairs (T1/T2, T2/T5, T3/T6, T4/T6). Higher correlations between FFR pitch contours would indicate a greater degree of merging in the brainstem. For evaluating cortical correlates of tone merging, we administered P300 in an oddball paradigm using blocks of relevant tone pairs (T1/T2, T2/T5, T3/T6, T4/T6) in adult native Cantonese speakers (n = 25) and analyzed the amplitude and latency. P300 amplitude/latency depends on the amount of effort a listener devotes to a task (Luck, 2005). In the current study, merging pairs are expected to be more effortful than non-merging pairs resulting in higher P300 amplitude and latency. Both brainstem and cortical measures of tone merging were analyzed alongside behavioral discrimination scores obtained from all the participants in order to understand the individual variability in neurophysiology contributing to sound change.

Materials and Methods

Two experiments were carried out in the current study. The first explored the subcortical representation of tone merging using FFR, and the second examined cortical representation of tone merging using P300. Additionally, behavioral data were collected in both experiments using AX discrimination tasks.

Stimuli for Neurophysiological Data Acquisition

A 25-year old male native speaker of Cantonese, phonetically-trained for production of Cantonese lexical tones, produced the syllable /ji/ with six Cantonese lexical tones making six unique words (Liu et al., 2015): /ji1/ “doctor,” /ji2/ “chair,” /ji3/ “meaning,” /ji4/ “son,” /ji5/ “ear” and /ji6/ “justice.” Figure 1 shows the contour of the six lexical tones of Cantonese (F0 ranges: 135–146, 105–134, 120–124, 85–99, 98–113, 98–106). Recordings were conducted in an acoustic booth using a Shure SM10A microphone and Praat (Boersma and Weenink, 2010) with a sampling rate of 44.1 kHz. Five versions of the stimuli were created with durations normalized to 150, 175, 200, 225, and 250 msec. A speech identification task was administered on 12 native Cantonese listeners to select the stimuli with the most appropriate stimulus duration for the experiment. The 175 ms stimuli were consistently and correctly identified and thus were used as stimuli for the experiment.

FIGURE 1
www.frontiersin.org

Figure 1. F0 contours of all six Cantonese tones (F0 ranges: T1:135-146 Hz, T2: 105-134 Hz, T3: 120-124 Hz, T4: 85-99 Hz, T5: 98-113 Hz, T6: 98-106 Hz, respectively).

Participants

Thirty native speakers of Cantonese (M age = 21.7 years; range = 18–29 years; 10 males), studying at the Chinese University of Hong Kong, participated in the FFR experiment. Of these, seventeen had some formal musical training (M years of training = 3.1; range = 1–13 years). Twenty-five participants (M age = 23.0; range = 21–27 years; 6 males), studying at the Chinese University of Hong Kong participated in the P300 experiment. Of these, sixteen had some formal musical training (M years of training = 3.2; range = 1–12 years). All participants had peripheral hearing sensitivity within 25 dB HL for 0.5–4 kHz. Written informed consents were obtained from all participants prior to the experiments. The Joint Chinese University of Hong Kong—New Territories East Cluster Clinical Research Ethics Committee approved the study.

Procedure

Experiment 1

FFR Acquisition

Brainstem responses collected from 3000 sweeps of each stimulus presented in alternating polarity were added to minimize stimulus artifacts and cochlear microphonics (Gorga et al., 1985; Skoe and Kraus, 2010). Stimuli were routed to the right ear through ear inserts (Compumedics 10 Ω) at 81 dB SPL with an interstimulus interval (offset to onset) jittered between 74 and 104 msec (Wong et al., 2007; Liu et al., 2015) using the Audio CPT module of STIM2 (Compumedics, USA). The order of stimulus presentation was randomized across participants. Responses were differentially collected from Ag/AgCl electrodes at Cz referenced to linked M1 and M2 (linked mastoids), with lower forehead as ground (Liu et al., 2015), using Synamps RT amplifier connected to Curry 7.05 workstation (Compumedics, El Paso, TX). Inter-electrode impedances were maintained at less than or equal to 1 kΩ. The responses were collected at a sampling rate of 20 kHz. Offline analysis consisting of artifact rejection, filtering, epoching, averaging, converting to wave files was done using Curry 7.05 (Compumedics, El Paso, TX). Sweeps with voltage beyond ± 35 μV were considered artifacts and rejected. The waveforms were band-pass filtered from 80 to 5000 Hz (with 6 dB roll off) with a 275 msec time window consisting of 50 msec pre-and post-stimulus baselines. The EEG recordings with more than 10% of the sweeps rejected (i.e., >300 rejections), were not included in further analyses.

FFR analysis procedures

In order to test phase-locking ability of the auditory brainstem, FFRs were band-pass filtered offline from 80 to 2500 Hz. Pitch contours of the FFRs were obtained using a periodicity detection autocorrelation algorithm in MATLAB 8.3 (The MathWorks, Inc, Natick, MA, USA). The FFR waveforms were converted from the temporal domain to the spectral domain and amplitude peaks around the F0 were extracted from seven non-overlapping 25-ms bins (0–25, 26–50, 51–75, 76–100, 101–125, 126–150, 151–175) (Krishnan et al., 2005). The F0 pitch contour of the FFR was constructed by connecting these peaks extracted from each time bin (Figure 2). Supplementary Figure 1 shows averaged FFR pitch contours from all subjects. The FFR pitch contours were used to calculate response-to-response correlation for each subject, which is a Pearson correlation between pitch contours (F0) of two brainstem responses. Here, response-to-response correlation serves as an indication of perceptual merging of the tone categories in brainstem and was calculated for the following tone pairs: T2/T5 (full-merger), T4/T6 (near-merger), T3/T6 (quasi-merger), and T1/T2 (non-merger). Strong positive response-to-response correlation values reflect a greater degree of merging for lexical tone pairs in the brainstem, whereas weak and/or negative correlations represent the absence of tone merging. Supplementary Tables 1, 2 provide correlation matrices for stimulus-to-stimulus and response-to-response correlations, respectively, for all possible tone-pair combinations.

FIGURE 2
www.frontiersin.org

Figure 2. Illustration of the method of extraction of F0 contour from FFR of a Cantonese lexical Tone 2 (T2) stimulus: (A) the FFRs in the time-domain are divided into 7 bins of 25 ms each; (B) Autocorrelation periodicity detection algorithm works on each bin separately; (C) F0 points are obtained at each bin that are connected to form the F0 contour of the FFR.

Behavioral Testing and Analyses

An AX discrimination task was used to test the four tone pairs (T2/T5, T4/T6, T3/T6, T1/T2). To enhance ecological validity, the speech discrimination task used words different from those used in the FFR experiment. These words were “tong” (T2/T5), “jy” (T4/T6), “bei” (T3/T6), and “min” (T1/T2), selected from a subset of stimuli from Wong et al. (2009). The stimuli were normalized for duration and intensity, and differed only in their pitch contours (Figure 3). Four practice trials (with 5 repetitions), different from the experimental items were provided to the participants to familiarize them with the task. Participants were presented with a total of 80 stimulus pairs (4 tone pairs × 5 repetitions × 2 “identical” sequences × 2 “different” sequences) with an interstimulus interval of 500 msec. The participants were required to press the appropriate button to indicate “same” or “different” for the pitch patterns of the tone pairs. For each lexical tone pair, discrimination scores were recorded, and d′ scores were further calculated by subtracting z-scores of hits from false alarms (d′ = ZFA–ZHit). This measures the subjects' discrimination sensitivity while accounting for response bias.

FIGURE 3
www.frontiersin.org

Figure 3. Time-normalized F0 contours (in Hz) of the four tone pairs used in the AX discrimination task: (A) bei_T6 [鼻,“nose”] – bei_T3 [臂,“arm”]; (B) jy_T5 [雨,“rain”] –jy_T2 [鱼,“fish”]; (C) min_T4 [綿,“cotton”] – min_T6 [麵,“noodle”]; (D) tong_T2 [糖,“sweet (candy)”] – tong_T1 [湯,“soup”]. Note: T1, T2, T3, T4, T5, and T6 correspond to the Cantonese high-level (Tone 1), high rising (Tone 2), mid-level (Tone 3), low-falling (Tone 4), low-rising (Tone 5), and low-level (Tone 6) tones, respectively.

Experiment 2

P300 Acquisition and Analysis

P300 data were collected for the four tone pairs: Tone 2/Tone 5 (T2/T5; full-merger), Tone 3/Tone 6 (T3/T6; quasi-merger), Tone 4/Tone 6 (T4/T6; near-merger), Tone 1/Tone 2 (T1/T2; non-merger). For P300 acquisition, we collected averaged EEG responses from 1000 sweeps of each pair of stimuli in an oddball paradigm (Standard: Rare: 80: 20). The order of stimuli (tone pairs) and their presentation (standard and rare stimuli) were counterbalanced across subjects. The stimuli were presented at an intensity of 80 dB SPL to the right ear via insert ER-3A insert earphones (Etymotic research) at a repetition rate of every 0.85 s. Participants were instructed to press the response button as soon as they heard a deviant stimulus. As we were interested in only looking at the temporal aspects, EEG data were collected with a 4-electrode Cz-(M1+M2)-ground montage using an Intelligent Hearing Systems unit (Miami, USA). The responses were filtered from 1 to 30 Hz, rejected for artifacts beyond ± 75 μV, baseline corrected and averaged over a time window of −100 msec (pre-stim baseline) to 500 ms. The P300 peaks were identified as the maximum peak amplitude on the waveforms elicited by the deviant stimuli in the latency region of 250 to 400 ms. The peaks were identified independently by two experienced EEG researchers and in case of a disagreement, opinion of a third rater was sought.

Behavioral Data Collection

Given the technical constraints of the EEG instrument used, the behavioral responses could not be recorded simultaneously and thus, were collected separately. In order to test the discrimination ability of the participants for the 4 tone pairs (T2/T5, T4/T6, T3/T6, T1/T2) used in the P300 task, an AX discrimination task was constructed using the same set of stimuli used in P300 recording (/ji1/ to /ji6/). Participants were presented with a total of 48 stimuli pairs (4 tone pairs × 3 repetitions × 2 “identical” sequence × 2 “different” sequence) with an interstimulus interval of 500 ms. The participants were required to press the appropriate button to indicate “same” or “different” for the pitches of the tone pairs. For each lexical tone pair, accuracy, d′ (ZFA–ZHit) and reaction time were calculated.

Results

Experiment 1: Brainstem and Behavior

Behavioral Results

The pattern of results on d′ revealed that the subjects best discriminated the non-merger pair (T1/T2) followed by the quasi-merger pair (T3/T6), near-merger pair (T4/T6) and full-merger pair (T2/T5) (Figure 4A). There was a main effect of category on d′ with a Greenhouse-Geisser adjustment to the degrees of freedom, F(2.28, 66.18) = 19.01, p <.001, ηp2 = 0.40. Further, post-hoc paired t-tests revealed that participants performed better on the near-merger than full-merger pair, t(29) = −4.33, p < 0.001, and better on the non-merger pair than the quasi-merger pair, t(29) = 3.61, p = 0.001. However, there was no significant difference between the d′ of the near-merger and quasi-merger pairs, t(29) = 0.69, p = 0.498.

FIGURE 4
www.frontiersin.org

Figure 4. (A) Discrimination of tone pairs in an AX discrimination task and; (B) Response-to-response correlation of FFRs of tone pairs varying in their degree of tone category merging. Full-merger pair (T2/T5) showed the lowest discrimination scores and highest response-to-response correlation, whereas non-merger pair (T1/T2) showed the highest discrimination scores and lowest response-to-response correlation (**p < 0.01, *p = 0.021; Error bars = ± SEM).

Brainstem Electrophysiology Findings

FFRs were obtained for all six lexical tones of Cantonese. Figure 5 depicts waveforms from the six Cantonese lexical tones (left panel) and their grand-averaged FFRs (right panel). In order to obtain a brainstem representation of perceptual merging, response-to-response correlations were calculated for T1/T2 (non-merger), T2/T5 (full-merger), T3/T6 (quasi-merger), and T4/T6 (near-merger) tone pairs. The full-merger pair (T2/T5) showed the strongest response-to-response correlation, followed by the near-merger (T4/T6), quasi-merger (T3/T6) and non-merger (T1/T2) pairs in a sharp gradation (Figure 4B). A main effect of tone category was obtained with a Greenhouse-Geisser adjustment, F(2.39, 69.3) = 40.09, p < 0.001, ηp2 = 0.58. Further, post-hoc paired t-tests revealed that there was a significant difference between the full-merger pair and the near-merger pair, t(29) = −3.41, p = 0.002; the quasi-merger pair and the non-merger pair, t(29) = −4.52, p < 0.001, with the near- and quasi-merger pairs not reaching significance after correction for multiple comparisons, t(29) = −2.43, p = 0.021.

FIGURE 5
www.frontiersin.org

Figure 5. Cantonese lexical tone stimuli (left panel) and coresponding grandaveraged FFRs (right panel). X-axis: Time (ms), Y-axis: Amplitude (Arbitrary units for stimuli in the left panel; μV for FFRs in the right panel). Initial and final 50 msec represents pre- and post-stimulus baselines.

Relationship between Behavioral and FFR Findings

In order to determine the relationship between behavioral discrimination abilities and brainstem measures, we calculated Pearson correlations between behavioral data (d′) of tone pairs and the corresponding response-to-response measures (Figure 6). For the non-merger (T1/T2) pair, there was no significant correlation with d′ (r = −0.11, p = 0.55). However, moderate correlations were observed between d′ and neurophysiological measures for full-merger (T2/T5) pair (d′: r = −0.59, p = 0.001), T3/T6 pair (d′: r = −0.57, p = 0.001), and near-merger (T4/T6) pair (d′: r = −0.44, p = 0.014). Musical experience did not correlate with d′ (T2/T5: r = 0.07, p = 0.722; T3/T6: r = 0.08, p = 0.678; T4/T6: r = 0.10, p = 0.593; T1/T2: r = 0.15, p = 0.435) and response-to response correlation (T2/T5: r = 0.05, p = 0.806; T3/T6: r = 0.07, p = 0.699; T4/T6: r = 0, p = 0.987; T1/T2: r = −0.14, p = 0.473).

FIGURE 6
www.frontiersin.org

Figure 6. Depicting the distribution of participants across response-to-response correlation and d′ for the four tone pair combinations. Non-merger pair (D) showed no significant correlation whereas other tone pairs (A–C) showed significant correlations with behavioral discrimination measures.

Experiment 2: Cortex and Behavior

Behavioral Results

There was a main effect of the tone categories on discrimination abilities (d′), F(1.81, 43.43) = 14.33, p < 0.001, ηp2 = 0.37. Further, post-hoc paired t-tests revealed a significant difference between the full-merger and near-merger pairs, t(24) = −3.65, p = 0.001, but no significant difference between the non-merger and quasi-merger pairs, t(24) = 0, p = 1, and quasi- and near-merger categories, t(24) = −1.98, p = 0.059 (Figure 7A).

FIGURE 7
www.frontiersin.org

Figure 7. Tone pairs varying in their degree of tone category merging (**p <0.01, *p <0.05, p = 0.022; Error bars = ± SEM). (A) Non-merger pair (T1/T2) was discriminated best, whereas full-merger pair (T2/T5) was discriminated worst; (B–D) Full-merger pair (T2/T5) showed highest reaction time, P300 latency and amplitude, whereas non-merger pair (T1/T2) showed lowest values on these parameters.

Additionally, there was a main effect of tone-categories on reaction time for discrimination, F(1.95, 46.88) = 4.10, p = 0.024, ηp2 = 0.15. Further, post-hoc paired t-tests revealed a significant difference between the full-merger and near-merger pairs, t(24) = 2.84, p = 0.009, but no significant difference between the non-merger and quasi-merger pairs, t(24) = 0.52, p = 0.607 or the quasi-merger and near merger pairs, t(24) = −0.41, p = 0.683 (Figure 7B).

Cortical Electrophysiology Findings

There was a main effect of the tone categories on P300 peak latency, F(2.29, 55.08) = 7.05, p = 0.001. Further, post-hoc paired t-tests revealed a significant difference between the quasi- and near-merger pairs, t(24) = −3.11, p = 0.005, and no significant difference between the near- and full-merger pairs, t(24) = −2.01, p = 0.056 and non- and quasi-merger pairs, t(24) = 0.29, p = 0.773 (Figures 7C, 8). Similarly, there was a main effect of the tone categories on P300 peak amplitude with a Greenhouse-Geisser adjustment, F(3, 67.22) = 3.21, p = 0.031, ηp2 = 0.51. Further, post-hoc paired t-tests revealed that the difference between the tone pairs did not reach significance after correction for multiple comparisons (full- and near-merger: t(24) = −2.45, p = 0.022; non- and quasi-merger, t(24) = −0.96, p = 0.345; quasi- and near-merger, t(24) = 0.85, p = 0.404; Figures 7D, 8).

FIGURE 8
www.frontiersin.org

Figure 8. Grand-averaged P300 waveforms for (A) Full-merger (T2/T5) (B) Near-merger (T4/T6) (C) Quasi-merger (T3/T6) (D) Non-merger (T1/T2). Full-merger pair elicited the highest P300 amplitude and longest latency.

Relationship between Behavioral and P300 Findings

P300 latency correlated significantly with the behavior (d′) only for the full-merger pair (r = −0.52, p = 0.008; Figure 9) but not with other tone pairs (quasi-merger: r = −0.09, p = 0.645; near-merger: r = 0.09, p = 0.668; non-merger: r = 0.26, p = 0.219). Additionally, no significant correlation between P300 amplitude and d′ were found (full-merger: r = −0.19, p = 0.364; quasi-merger: r = 0.048, p = 0.819; near-merger: r = −0.01, p = 0.95; non-merger: r = 0.25, p = 0.23). Musical experience did not correlate with d′ (T2/T5: r = 0.06, p = 0.766; T3/T6: r = −0.09, p = 0.684; T4/T6: r = −0.11, p = 0.603; T1/T2: r = −0.21, p = 0.331), P300 amplitude (T2/T5: r = −0.11, p = 0.595; T3/T6: r = 0.24, p = 0.255; T4/T6: r = −0.37, p = 0.071; T1/T2: r = −0.05, p = 0.807) and P300 latency (T2/T5: r = 0.05, p = 0.823; T3/T6: r = −0.08, p = 0.707; T4/T6: r = −0.08, p = 0.698; T1/T2: r = −0.21, p = 0.314).

FIGURE 9
www.frontiersin.org

Figure 9. Full-merger pair (T2/T5) showed a significant correlation (r = −0.52, p = 0.008) between d′ and P300 latency.

Discussion

In the current study, we explored sound change from behavioral and neurophysiological perspectives. Given that the Hong Kong Cantonese tone system is undergoing merging, we used it as a vehicle to study sound change. We found that there was a gradation of degree of merging across the tone categories. On average, Cantonese participants showed poorest discrimination or maximum merging for the full-merger pair (T2/T5) and best discrimination or least merging for the non-merger pair (T1/T2), while the quasi- (T3/T6) and near-merger (T4/T6) pairs showed comparable discrimination and merging. Neurophysiological measures including cortical and subcortical responses revealed a similar gradation of the tone categories. However, subcortical responses (FFR) revealed a sharper gradation or categorization than cortical neurophysiological responses and behavioral discrimination scores. Additionally, there were more individual differences in the subcortical and cortical processes for tone pairs with a greater degree of merging relative to those with less degree of merging. We propose that these individual differences can be a contributory factor to sound change.

The behavioral discrimination data are compatible with the idea that perceptual similarity/confusability leads to sound change (Ohala, 1981, 2012). Based on reaction time data in an AX discrimination paradigm, (Mok et al., 2013) found that their participants took longer on T2/T5 discrimination than T1/T2 discrimination. Recently, there have been studies (Law et al., 2013) classifying participants into merger and/or non-merger groups based on their speech production and/or perception. The literature on Cantonese tone merging has been limited to studying merging as a discrete phenomenon. However, in order to systematically investigate sound change in Cantonese, tone merging must be viewed as a continuum across the tone categories. In the present study, we evaluated one non-merging pair and three merging pairs of Cantonese lexical tones that varied in their degrees of merging. The current study, for the first time investigated the sound change as a continuum of Cantonese tone merging.

Individual variability is considered to be an important factor for dissemination of variants of sound change (Witkin et al., 1975; Messick, 1976; Ausburn and Ausburn, 1978; Yu, 2010, 2013). In the current study, variability in speech perception across individuals was established behaviorally and neurophysiologically. Significant correlations were found between behavioral discrimination scores and FFR response-to-response correlations for the tone merging pairs, whereas there was no correlation between the behavioral and neurophysiological aspects of the non-merger pair. Similarly, there was a significant correlation between the behavioral discrimination scores for the full-merger pair and corresponding P300 latency, whereas there were no significant correlations in the lesser or non-merging pairs. The absence of reliable correlations between behavioral and neurophysiological measures for the lesser or non-merging pairs and the presence of reliable correlations for the merging pairs confirms that the individual variability among listeners is equally well-represented in their neurophysiology and behavior, and may in turn contribute to sound change. Thus, the findings from the present study establish a plausible link between sound change, behavior and neurophysiology.

In the FFR data, stronger response-to-response correlations suggested more merging of tone categories in the brainstem. In the cortical neurophysiological data, higher values of amplitude and latency corresponded with increased merging of tones in the tone pairs. A sharper gradation of the merging phenomenon was evident in the brainstem as compared to cortical and behavioral data, probably because of sharp phase locking abilities of the brainstem. Recently, Law et al. (2013) conducted an ERP study exploring tone merging at cortical levels using MMN and P3a. They found that individuals who performed poorly on T4/T6 (near-merger) discrimination tasks had a smaller MMN and P3a amplitude. However, as merger pairs with varying degrees of merging were not included, their study was limited to investigating merging as a discrete phenomenon and not as a continuum to account for sound change. In the current study, instead of looking at inattentive levels of processing using MMN or P3a, we looked at attentive levels using P300. Additionally, a P300 task challenges the capacities of the auditory and attention system as a whole and reflects the effort devoted by individuals in discriminating stimuli (Luck, 2005). Therefore, use of a P300 task, as in the current study, is informative of active language processing and a better indicator of tone merging.

The source of individual variability in the behavioral and neurophysiological findings of the current study can be discussed in light of bottom-up processing (Krishnan et al., 2005, 2009, 2010), top-down processing (Suga, 2008; Parbery-Clark et al., 2009; Krizman et al., 2012; Song et al., 2012; Hairston et al., 2013; Skoe and Kraus, 2013), and the predictive tuning hypothesis (Chandrasekaran et al., 2014). The brainstem has been found to be sensitive to language experience as lexical tones are more accurately represented in tone language speakers than non-tone language speakers. Krishnan et al. (2005, 2009, 2010) argue that the experience-dependent effects in brainstem responses are due to local reorganization of brainstem circuits via excitatory and inhibitory synaptic plasticity. This account has been supported by animal models where it was found that auditory midbrain neurons rapidly adapt to the dynamic stimulus characteristics (Escabí et al., 2003; Dean et al., 2005; Pérez-González et al., 2005; Dahmen et al., 2010). In other words, the auditory brainstem is locally malleable. In the current study, the participants who may be inducing variability (called innovators) may be the ones with aberrant encoding of lexical tones at the level of the inferior colliculi. As a result of this, there is an inefficient transmission of information to cortex that is leading to deficient perception.

Recently, it has also been found that the effects visible at the level of the brainstem originate at the level of the cortex and are driven via corticofugal pathways from the cortex to the brainstem (Suga, 2008; Parbery-Clark et al., 2009; Bajo et al., 2010; Krizman et al., 2012; Song et al., 2012; Hairston et al., 2013; Skoe and Kraus, 2013). This view has gained support from animal models where inactivation of auditory cortex has led to disruption of brainstem plasticity (Suga et al., 2000, 2002; Zhou and Jen, 2000; Suga, 2008). Gao and Suga (1998) found that behaviorally relevant sounds reflect more activity in the inferior colliculi as compared to acoustic stimulation alone. Tone merging originating in the cortex, as observed in the current study, could have driven merging at the level of the brainstem. We found that the tone pairs showed similar trends of merging in brainstem and cortical measures. The pairs with longer P300 latencies also showed stronger response-to-response correlations, and vice versa. These findings show that the individuals inducing variability might have learned the pitch contours of the lexical tones in a divergent manner. Learning could have led to divergent cortical changes. Corticofugal effects of this divergent experience-dependent learning could have led to deviant representations in the brainstem. Though, both cortical and brainstem mechanisms may be affected with divergent learning, the effects are more well-defined and represented in the brainstem. This could be due to more accurate phase locking and fidelity of representation of lexical tones in the brainstem.

The findings of the current study can also be explained using the predictive tuning model (Chandrasekaran et al., 2014) which proposes that there is continuous, online modulation of brainstem encoding by the auditory cortex via corticofugal pathways (Chandrasekaran et al., 2009a; Chandrasekaran and Kraus, 2010) along with local processes in the inferior colliculi still being active. Signal representation gets enhanced at each level when the stimulus matches the expectation or prediction from a higher level. In the current study, it is possible that individuals contributing to variability (innovators) are not able to predict the F0 contour of the lexical tones accurately and thus are matching to a slightly deviant representation at the cortical levels as a result of which there is enhancement of slightly deviant representation at the brainstem level.

Though the results of the current study provide an insight into the subcortical and cortical neural representations of sound change in a group of Cantonese listeners, determining a causal relationship between the two is beyond the scope of this study. The results from the current study have implications toward studying atypical populations (e.g., autism) with cognitive and/or perceptual aspects deviant to the typical population, in order to investigate the neurophysiological indicators of sound change in greater detail. Future studies could use the latest pitch-specific ERP techniques such as cortical pitch response (Krishnan et al., 2012, 2014, 2015, 2016) to probe sound change in tone languages. Further, as other linguistic factors such as frequency of occurrence have been known to affect sound change (Hooper, 1976; Pierrehumbert, 2001; Abramowicz, 2007; Zhao and Jurafsky, 2009), future studies may conduct an in-depth investigation of neurophysiological indicators of sound change from the perspective of frequency of usage and exemplar theory (Abramowicz, 2007). Concomitantly, the data from the current study also exhibited a lack of influence of musical training on behavioral and neurophysiological differentiation of lexical tone contrasts. For investigating the key mechanisms responsible for these findings, another study from our group is underway.

Conclusions

In the current study, we investigated the contribution of individual variability toward sound change in Hong Kong Cantonese. We found that the sound change in Cantonese, indicated by reshaping of tone categories was observable at the levels of the brainstem and cortex. Moreover, we found that the newly merging categories that are not yet clearly expressed at the behavioral and cortical levels appear delineated at the subcortical level. These findings have implications toward the active contribution of the brainstem in driving sound change. Overall, we speculate that the individuals with divergent encoding of speech sounds in their subcortical and cortical processes might produce speech in a deviant manner that spreads across the community and leads to sound change over a period of time. However, a causal relationship between the neural representations and sound change still remains an open question.

Ethics Statement

The Joint Chinese University of Hong Kong—New Territories East Cluster Clinical Research Ethics Committee. Written informed consents were obtained from all participants prior to the experiments. The Joint Chinese University of Hong Kong—New Territories East Cluster Clinical Research Ethics Committee approved the study.

Author Contributions

AM: Data collection, Experiment designing (Behavioral and EEG), Data analysis, Manuscript writing. FL: Experiment designing (Behavioral), Data collection and analysis. MA: Data analysis, Manuscript writing. PW: Experiment designing (Behavioral and EEG), Data analysis, Manuscript writing.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

This research was supported by Dr. Stanley Ho Medical Development Foundation, U.S. National Institutes of Health Grant Nos. R01DC008333 and R01DC013315, National Science Foundation Grant No. BCS-1125144, Research Grants Council of Hong Kong Grant Nos. 477513 and 14117514 to PW and Australian Research Council Discovery Early Career Research Award DE150101053 to MA. We also thank Doris Lau and Joseph Lau for their assistance in this study.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/article/10.3389/fnhum.2016.00652/full#supplementary-material

References

Abramowicz, L. (2007). Sociolinguistics meets exemplar theory: Frequency and recency effects in (ing). Univ. Pa. Work. Pap. Linguist. 13, 3.

Google Scholar

Anderson, S., Parbery-Clark, A., White-Schwoch, T., Drehobl, S., and Kraus, N. (2013). Effects of hearing loss on the subcortical representation of speech cues. J. Acoust. Soc. Am. 133, 3030–3038. doi: 10.1121/1.4799804

PubMed Abstract | CrossRef Full Text | Google Scholar

Ausburn, L. J., and Ausburn, F. B. (1978). Cognitive styles: some information and implications for instructional design. ECTJ 26, 337–354.

Google Scholar

Bajo, V. M., Nodal, F. R., Moore, D. R., and King, A. J. (2010). The descending corticocollicular pathway mediates learning-induced auditory plasticity. Nat. Neurosci. 13, 253–260. doi: 10.1038/nn.2466

PubMed Abstract | CrossRef Full Text | Google Scholar

Beckner, C., Blythe, R., Bybee, J., Christiansen, M. H., Croft, W., Ellis, N. C., et al. (2009). Language is a complex adaptive system: position paper. Lang. Learn. 59, 1–26. doi: 10.1111/j.1467-9922.2009.00533.x

CrossRef Full Text | Google Scholar

Bidelman, G. M., Gandour, J. T., and Krishnan, A. (2009). Cross-domain effects of music and language experience on the representation of pitch in the human auditory brainstem. J. Cogn. Neurosci. 23, 425–434. doi: 10.1162/jocn.2009.21362

PubMed Abstract | CrossRef Full Text | Google Scholar

Bloomfield, L. (1933). Language. Chicago: University of Chicago Press.

Google Scholar

Boersma, P., and Weenink, D (2010). {P}raat: Doing Phonetics by Computer. Available online at: http://www.praat.org (Accessed September 21, 2015)

Browman, C., and Goldstein, L. (2014). “Tiers in articulatory phonology, with some implications for casual speech,” in The Oxford Handbook of Language Production, eds M. Goldrick, V. Ferreira, and M. Miozzo (Cambridge: Oxford University Press), 341–376.

Buchwald, J. S., Guthrie, D., Schwafel, J., Erwin, R. J., and Van Lancker, D. (1994). Influence of language structure on brain-behavior development. Brain Lang. 46, 607–619. doi: 10.1006/brln.1994.1033

PubMed Abstract | CrossRef Full Text | Google Scholar

Chandrasekaran, B., and Kraus, N. (2010). The scalp-recorded brainstem response to speech: neural origins and plasticity. Psychophysiology 47, 236–246. doi: 10.1111/j.1469-8986.2009.00928.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Chandrasekaran, B., Krishnan, A., and Gandour, J. T. (2007). Mismatch negativity to pitch contours is influenced by language experience. Brain Res. 1128, 148–156. doi: 10.1016/j.brainres.2006.10.064

PubMed Abstract | CrossRef Full Text | Google Scholar

Chandrasekaran, B., Krishnan, A., and Gandour, J. T. (2009a). Relative influence of musical and linguistic experience on early cortical processing of pitch contours. Brain Lang. 108, 1–9. doi: 10.1016/j.bandl.2008.02.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Chandrasekaran, B., Krishnan, A., and Gandour, J. T. (2009b). Sensory processing of linguistic pitch as reflected by the mismatch negativity. Ear Hear. 30, 552–558. doi: 10.1097/AUD.0b013e3181a7e1c2

PubMed Abstract | CrossRef Full Text | Google Scholar

Chandrasekaran, B., Skoe, E., and Kraus, N. (2014). An integrative model of subcortical auditory plasticity. Brain Topogr. 27, 539–552. doi: 10.1007/s10548-013-0323-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Cheng, R. (1987). Borrowing and internal development in lexical change—a comparison of taiwanese words and their mandarin equivalents / 从台湾话和华语之间的对等语的比较看语汇变化的两个模式:移借和内部发展. J. Chin. Linguist. 15, 105–131.

Google Scholar

Dahmen, J. C., Keating, P., Nodal, F. R., Schulz, A. L., and King, A. J. (2010). Adaptation to stimulus statistics in the perception and neural representation of auditory space. Neuron 66, 937–948. doi: 10.1016/j.neuron.2010.05.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Dean, I., Harper, N. S., and McAlpine, D. (2005). Neural population coding of sound level adapts to stimulus statistics. Nat. Neurosci. 8, 1684–1689. doi: 10.1038/nn1541

PubMed Abstract | CrossRef Full Text | Google Scholar

Eckert, P. (2000). Language Variation as Social Practice: The Linguistic Construction of Identity in Belten High. Oxford: Wiley-Blackwell.

Escabí, M. A., Miller, L. M., Read, H. L., and Schreiner, C. E. (2003). Naturalistic auditory contrast improves spectrotemporal coding in the cat inferior colliculus. J. Neurosci. 23, 11489–11504.

Google Scholar

Frenck-Mestre, C., Meunier, C., Espesser, R., Daffner, K., and Holcomb, P. (2005). Perceiving nonnative vowels: the effect of context on perception as evidenced by event-related brain potentials. J. Speech Lang. Hear. Res. 48, 1496. doi: 10.1044/1092-4388(2005/104)

PubMed Abstract | CrossRef Full Text | Google Scholar

Fung, R., and Wong, S. (2011). “The acoustic analysis of the new rising tone in Hong Kong Cantonese,” in Proceedings of 17th International Congress of Phonetic Sciences (Hong Kong), 715–718.

Galbraith, G. C., Arbagey, P. W., Branski, R., Comerci, N., and Rector, P. M. (1995). Intelligible speech encoded in the human brain stem frequency-following response. Neuroreport 6, 2363–2367. doi: 10.1097/00001756-199511270-00021

PubMed Abstract | CrossRef Full Text | Google Scholar

Gao, E., and Suga, N. (1998). Experience-dependent corticofugal adjustment of midbrain frequency map in bat auditory system. Proc. Natl. Acad. Sci. U.S.A. 95, 12663–12670. doi: 10.1073/pnas.95.21.12663

PubMed Abstract | CrossRef Full Text | Google Scholar

Gauchat, L. (1905). L'unité Phonétique dans le patois d'une Commune. Aus Romanischen Sprachen und Literaturen: Festschrift Heinrich Mort. Halle: Max Niemeyer.

Gorga, M. P., Worthington, D. W., Reiland, J. K., Beauchaine, K. A., and Goldgar, D. E. (1985). Some comparisons between auditory brain stem response thresholds, latencies, and the pure-tone audiogram. Ear Hear. 6, 105–112. doi: 10.1097/00003446-198503000-00008

PubMed Abstract | CrossRef Full Text | Google Scholar

Haeri, N. (1996). The Sociolinguistic Market of Cairo: Gender, Class, and Education. London; New York, NY: Kegan Paul International; Distributed by Columbia University Press.

Google Scholar

Hairston, W. D., Letowski, T. R., and McDowell, K. (2013). Task-related suppression of the brainstem frequency following response. PLoS ONE 8:e55215. doi: 10.1371/journal.pone.0055215

PubMed Abstract | CrossRef Full Text | Google Scholar

Hermann, E. (1929). Lautveränderungen in der Individualsprache einer Mundart. Nachrichten der Gesellschaft der Wissenschaften zu Goettingen Phl.-his. Klasse 9, 195–214.

Hooper, J. B. (1976). “Word frequency in lexical diffusion and the source of morphophonological change,” in Current Progress in Historical Linguistics, ed W. M. Christie (Amsterdam: North-Holland), 96–105.

Hornickel, J., Skoe, E., and Kraus, N. (2009). Subcortical laterality of speech encoding. Audiol. Neurotol. 14, 198–207. doi: 10.1159/000188533

PubMed Abstract | CrossRef Full Text | Google Scholar

Krishnan, A., Bidelman, G. M., Smalt, C. J., Ananthakrishnan, S., and Gandour, J. T. (2012). Relationship between brainstem, cortical and behavioral measures relevant to pitch salience in humans. Neuropsychologia 50, 2849–2859. doi: 10.1016/j.neuropsychologia.2012.08.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Krishnan, A., Gandour, J. T., Ananthakrishnan, S., and Vijayaraghavan, V. (2015). Language experience enhances early cortical pitch-dependent responses. J. Neurolinguist. 33, 128–148. doi: 10.1016/j.jneuroling.2014.08.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Krishnan, A., Gandour, J. T., and Bidelman, G. M. (2010). The effects of tone language experience on pitch processing in the brainstem. J. Neurolinguist. 23, 81–95. doi: 10.1016/j.jneuroling.2009.09.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Krishnan, A., Gandour, J. T., Bidelman, G. M., and Swaminathan, J. (2009). Experience dependent neural representation of dynamic pitch in the brainstem. Neuroreport 20, 408–413. doi: 10.1097/WNR.0b013e3283263000

PubMed Abstract | CrossRef Full Text | Google Scholar

Krishnan, A., Gandour, J. T., and Suresh, C. H. (2014). Cortical pitch response components show differential sensitivity to native and nonnative pitch contours. Brain Lang. 138, 51–60. doi: 10.1016/j.bandl.2014.09.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Krishnan, A., Gandour, J. T., and Suresh, C. H. (2016). Language-experience plasticity in neural representation of changes in pitch salience. Brain Res. 1637, 102–117. doi: 10.1016/j.brainres.2016.02.021

PubMed Abstract | CrossRef Full Text | Google Scholar

Krishnan, A., Xu, Y., Gandour, J., and Cariani, P. (2005). Encoding of pitch in the human brainstem is sensitive to language experience. Cogn. Brain Res. 25, 161–168. doi: 10.1016/j.cogbrainres.2005.05.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Krizman, J., Marian, V., Shook, A., Skoe, E., and Kraus, N. (2012). Subcortical encoding of sound is enhanced in bilinguals and relates to executive function advantages. Proc. Natl. Acad. Sci. U.S.A. 109, 7877–7881. doi: 10.1073/pnas.1201575109

PubMed Abstract | CrossRef Full Text | Google Scholar

Labov, W. (1980). “The social origins of sound change,” in Locating Language in Time and Space, ed W. Labov (New York, NY: Academic Press), 251–266.

Google Scholar

Labov, W. (1990). The intersection of sex and social class in the course of linguistic change. Lang. Var. Change 2, 205–254. doi: 10.1017/S0954394500000338

CrossRef Full Text

Labov, W. (2001). Principles of Linguistic Change, Social Factors (Vol. 2) Hardcover. Oxford: Wiley-Blackwell.

Google Scholar

Labov, W. (2006). The Social Stratification of English in New York City. Cambridge: Cambridge University Press.

Google Scholar

Labov, W. (2010). Principles of Linguistic Change, Cognitive and Cultural Factors (Vol. 3) Paperback. Oxford: Wiley-Blackwell.

Google Scholar

Labov, W., Ash, S., and Boberg, C. (2005). The Atlas of North American English: Phonetics, Phonology and Sound Change. Berlin: Mouton de Gruyter.

Google Scholar

Law, S.-P., Fung, R., and Kung, C. (2013). An ERP study of good production visà-vis poor perception of tones in Cantonese: implications for top-down speech processing. PLoS ONE 8:e54396. doi: 10.1371/journal.pone.0054396

CrossRef Full Text | Google Scholar

Liu, F., Maggu, A. R., Lau, J. C. Y., and Wong, P. C. M. (2015). Brainstem encoding of speech and musical stimuli in congenital amusia: evidence from Cantonese speakers. Front. Hum. Neurosci. 8:1029. doi: 10.3389/fnhum.2014.01029

PubMed Abstract | CrossRef Full Text | Google Scholar

Luck, S. (2005). An Introduction to the Event-Related Potential Technique (Cognitive Neuroscience). A Bradford Book. Available online at: http://www.amazon.ca/exec/obidos/redirect?tag=citeulike09-20&path=ASIN/0262621967 (Accessed September 21, 2015).

Maiste, A. C., Wiens, A. S., Hunt, M. J., Scherg, M., and Picton, T. W. (1995). Event-related potentials and the categorical perception of speech sounds. Ear Hear. 16, 68–90. doi: 10.1097/00003446-199502000-00006

PubMed Abstract | CrossRef Full Text | Google Scholar

Messick, S. (ed.). (1976). Individuality in Learning: Implications of Cognitive Styles and Creativity for Human Development, 1st Edn. San Francisco, CA: Jossey-Bass Inc.

Milroy, J., and Milroy, L. (1985). Linguistic change, social network and speaker innovation. J. Linguist. 21, 339–384. doi: 10.1017/S0022226700010306

CrossRef Full Text | Google Scholar

Mok, P. P., Zuo, D., and Wong, P. W. (2013). Production and perception of a sound change in progress: tone merging in Hong Kong Cantonese. Lang. Var. Change 25, 341–370. doi: 10.1017/S0954394513000161

CrossRef Full Text | Google Scholar

Musacchia, G., Sams, M., Skoe, E., and Kraus, N. (2007). Musicians have enhanced subcortical auditory and audiovisual processing of speech and music. Proc. Natl. Acad. Sci. U.S.A. 104, 15894–15898. doi: 10.1073/pnas.0701498104

PubMed Abstract | CrossRef Full Text | Google Scholar

Musacchia, G., Strait, D., and Kraus, N. (2008). Relationships between behavior, brainstem and cortical encoding of seen and heard speech in musicians and non-musicians. Hear. Res. 241, 34–42. doi: 10.1016/j.heares.2008.04.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Näätänen, R., Paavilainen, P., Rinne, T., and Alho, K. (2007). The mismatch negativity (MMN) in basic research of central auditory processing: a review. Clin. Neurophysiol. 118, 2544–2590. doi: 10.1016/j.clinph.2007.04.026

PubMed Abstract | CrossRef Full Text | Google Scholar

Ohala, J. J. (1981). “The listener as a source of language change,” in Papers from the Parasession on Language and Behavior, eds C. Masek, R. Hendrick, and M. Miller (Chicago: Chicago Linguistics Society), 178–203.

Ohala, J. J. (1990). What's Cognitive, What's Not in Sound Change. Duisburg: LAUD.

Google Scholar

Ohala, J. J. (1993). Sound change as nature's speech perception experiment. Speech Commun. 13, 155–161. doi: 10.1016/0167-6393(93)90067-U

CrossRef Full Text | Google Scholar

Ohala, J. J. (2012). “The origin of sound patterns in vocal tract constraints,” in The Production of Speech, ed P. F. MacNeilage (New York, NY: Springer Science & Business Media), 189–216.

Ou, J., Law, S.-P., and Fung, R. (2015). Relationship between individual differences in speech processing and cognitive functions. Psychon. Bull. Rev. 22, 1725–1732. doi: 10.3758/s13423-015-0839-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Parbery-Clark, A., Skoe, E., and Kraus, N. (2009). Musical experience limits the degradative effects of background noise on the neural processing of sound. J. Neurosci. 29, 14100–14107. doi: 10.1523/JNEUROSCI.3256-09.2009

PubMed Abstract | CrossRef Full Text | Google Scholar

Pérez-González, D., Malmierca, M. S., and Covey, E. (2005). Novelty detector neurons in the mammalian auditory midbrain. Eur. J. Neurosci. 22, 2879–2885. doi: 10.1111/j.1460-9568.2005.04472.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Pierrehumbert, J. B. (2001). “Exemplar dynamics: Word frequency, lenition and contrast,” in Typological Studies in Language, eds J. L. Bybee and P. J. Hopper (Amsterdam: John Benjamins Publishing Company), 137. Available online at: https://benjamins.com/catalog/tsl.45.08pie (Accessed August 15, 2016)

Polich, J. (2007). Updating P300: an integrative theory of P3a and P3b. Clin. Neurophysiol. Off. J. Int. Fed. Clin. Neurophysiol. 118, 2128–2148. doi: 10.1016/j.clinph.2007.04.019

PubMed Abstract | CrossRef Full Text | Google Scholar

Russo, N. M., Hornickel, J., Nicol, T., Zecker, S., and Kraus, N. (2010). Biological changes in auditory function following training in children with autism spectrum disorders. Behav. Brain Funct. 6:60. doi: 10.1186/1744-9081-6-60

PubMed Abstract | CrossRef Full Text | Google Scholar

Russo, N., Nicol, T., Trommer, B., Zecker, S., and Kraus, N. (2009). Brainstem transcription of speech is disrupted in children with autism spectrum disorders. Dev. Sci. 12, 557–567. doi: 10.1111/j.1467-7687.2008.00790.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Shen, Z. (1997). Exploring the dynamic aspect of sound change. J Chin. Linguist. Monogr. Ser. 11, 1–189.

Google Scholar

Skoe, E., and Kraus, N. (2010). Hearing it again and again: on-line subcortical plasticity in humans. PLOS ONE 5:e13645. doi: 10.1371/journal.pone.0013645

PubMed Abstract | CrossRef Full Text | Google Scholar

Skoe, E., and Kraus, N. (2013). Musical training heightens auditory brainstem function during sensitive periods in development. Front. Psychol. 4:622. doi: 10.3389/fpsyg.2013.00622

PubMed Abstract | CrossRef Full Text | Google Scholar

Song, J. H., Skoe, E., Banai, K., and Kraus, N. (2012). Training to improve hearing speech in noise: biological mechanisms. Cereb. Cortex 22, 1180–1190. doi: 10.1093/cercor/bhr196

PubMed Abstract | CrossRef Full Text | Google Scholar

Suga, N. (2008). Role of corticofugal feedback in hearing. J. Comp. Physiol. A 194, 169–183. doi: 10.1007/s00359-007-0274-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Suga, N., Gao, E., Zhang, Y., Ma, X., and Olsen, J. F. (2000). The corticofugal system for hearing: recent progress. Proc. Natl. Acad. Sci. U.S.A. 97, 11807–11814. doi: 10.1073/pnas.97.22.11807

PubMed Abstract | CrossRef Full Text | Google Scholar

Suga, N., Xiao, Z., Ma, X., and Ji, W. (2002). Plasticity and corticofugal modulation for hearing in adult animals. Neuron 36, 9–18. doi: 10.1016/S0896-6273(02)00933-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Swaminathan, J., Krishnan, A., and Gandour, J. T. (2008). Pitch encoding in speech and nonspeech contexts in the human auditory brainstem. Neuroreport 19, 1163–1167. doi: 10.1097/WNR.0b013e3283088d31

PubMed Abstract | CrossRef Full Text | Google Scholar

Trudgill, P. (1974). The Social Differentiation of English in Norwich. Cambridge: CUP Archive.

Google Scholar

Velde, H. V., de Sankoff, G., Blondeau, H., and Charity, A. (2001). “Individual roles in a real-time change: Montreal (r-r),” in 'r-atics : Sociolinguistic, Phonetic and Phonological Characteristics of /r/, eds H. V. de Velde and R. van Hout (Bruxelles: ILVP), 141–158.

Google Scholar

Weinberg, M. (1974). Un Aspecto Sociolinguistico del Espanol Bonaerense: La -S en Bahia Blanca. Bahia Blanca: Cuadernos de Linguisticca.

Wible, B., Nicol, T., and Kraus, N. (2005). Correlation between brainstem and cortical auditory processes in normal and language-impaired children. Brain 128, 417–423. doi: 10.1093/brain/awh367

PubMed Abstract | CrossRef Full Text | Google Scholar

Witkin, H. A., Moore, C. A., Goodenough, D. R., and Cox, P. W. (1975). Field-dependent and field-independent cognitive styles and their educational implications. ETS Res. Bull. Ser. 1975, 1–64. doi: 10.1002/j.2333-8504.1975.tb01065.x

CrossRef Full Text | Google Scholar

Wolfe, P. M. (1972). Linguistic Change and the Great Vowel Shift in English. Berkeley: University of California Press.

Google Scholar

Wong, A. M.-Y., Ciocca, V., and Yung, S. (2009). The perception of lexical tone contrasts in Cantonese children with and without specific language impairment (SLI). J. Speech Lang. Hear. Res. 52, 1493–1509. doi: 10.1044/1092-4388(2009/08-0170)

PubMed Abstract | CrossRef Full Text | Google Scholar

Wong, P. C. M., Skoe, E., Russo, N. M., Dees, T., and Kraus, N. (2007). Musical experience shapes human brainstem encoding of linguistic pitch patterns. Nat. Neurosci. 10, 420–422. doi: 10.1038/nn1872

PubMed Abstract | CrossRef Full Text | Google Scholar

Yu, A. C. L. (2010). Perceptual compensation is correlated with individuals' “Autistic” traits: implications for models of sound change. PLoS ONE 5:e11950. doi: 10.1371/journal.pone.0011950

PubMed Abstract | CrossRef Full Text | Google Scholar

Yu, A. C. L. (2013). Origins of Sound Change: Approaches to Phonologization, 1st Edn. Oxford: Oxford University Press.

Google Scholar

Zhao, Y., and Jurafsky, D. (2009). The effect of lexical frequency and Lombard reflex on tone hyperarticulation. J. Phon. 37, 231–247. doi: 10.1016/j.wocn.2009.03.002

CrossRef Full Text | Google Scholar

Zheng, H.-Y., Minett, J. W., Peng, G., and Wang, W. S. (2012). The impact of tone systems on the categorical perception of lexical tones: an event-related potentials study. Lang. Cogn. Process. 27, 184–209. doi: 10.1080/01690965.2010.520493

CrossRef Full Text | Google Scholar

Zhou, X., and Jen, P. H.-S. (2000). Brief and short-term corticofugal modulation of subcortical auditory responses in the big brown bat, Eptesicus fuscus. J. Neurophysiol. 84, 3083–3087.

PubMed Abstract | Google Scholar

Keywords: language change, phonetics, individual variability, brainstem encoding, neurophysiology

Citation: Maggu AR, Liu F, Antoniou M and Wong PCM (2016) Neural Correlates of Indicators of Sound Change in Cantonese: Evidence from Cortical and Subcortical Processes. Front. Hum. Neurosci. 10:652. doi: 10.3389/fnhum.2016.00652

Received: 19 June 2016; Accepted: 08 December 2016;
Published: 23 December 2016.

Edited by:

Mikhail Lebedev, Duke University, USA

Reviewed by:

Michael Walsh, University of Stuttgart, Germany
Yi Zheng, École Normale Supérieure, France

Copyright © 2016 Maggu, Liu, Antoniou and Wong. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Patrick C. M. Wong, p.wong@cuhk.edu.hk

Download