<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Hum. Neurosci.</journal-id>
<journal-title>Frontiers in Human Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Hum. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-5161</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnhum.2022.890065</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Inter-Trial Formant Variability in Speech Production Is Actively Controlled but Does Not Affect Subsequent Adaptation to a Predictable Formant Perturbation</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Wang</surname> <given-names>Hantao</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1707250/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Max</surname> <given-names>Ludo</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/160222/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Department of Speech and Hearing Sciences, University of Washington</institution>, <addr-line>Seattle, WA</addr-line>, <country>United States</country></aff>
<aff id="aff2"><sup>2</sup><institution>Haskins Laboratories</institution>, <addr-line>New Haven, CT</addr-line>, <country>United States</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Jeffery A. Jones, Wilfrid Laurier University, Canada</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Takemi Mochida, Nippon Telegraph and Telephone, Japan; David Jenson, Washington State University, United States; Ding-lan Tang, University of Wisconsin-Madison, United States</p></fn>
<corresp id="c001">&#x002A;Correspondence: Ludo Max, <email>LudoMax@uw.edu</email></corresp>
<fn fn-type="other" id="fn004"><p>This article was submitted to Speech and Language, a section of the journal Frontiers in Human Neuroscience</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>07</day>
<month>07</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2022</year>
</pub-date>
<volume>16</volume>
<elocation-id>890065</elocation-id>
<history>
<date date-type="received">
<day>05</day>
<month>03</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>14</day>
<month>06</month>
<year>2022</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2022 Wang and Max.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>Wang and Max</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>Despite ample evidence that speech production is associated with extensive trial-to-trial variability, it remains unclear whether this variability represents merely unwanted system noise or an actively regulated mechanism that is fundamental for maintaining and adapting accurate speech movements. Recent work on upper limb movements suggest that inter-trial variability may be not only actively regulated based on sensory feedback, but also provide a type of workspace exploration that facilitates sensorimotor learning. We therefore investigated whether experimentally reducing or magnifying inter-trial formant variability in the real-time auditory feedback during speech production (a) leads to adjustments in formant production variability that compensate for the manipulation, (b) changes the temporal structure of formant adjustments across productions, and (c) enhances learning in a subsequent adaptation task in which a predictable formant-shift perturbation is applied to the feedback signal. Results show that subjects gradually increased formant variability in their productions when hearing auditory feedback with reduced variability, but subsequent formant-shift adaptation was not affected by either reducing or magnifying the perceived variability. Thus, findings provide evidence for speakers&#x2019; active control of inter-trial formant variability based on auditory feedback from previous trials, but&#x2013;at least for the current short-term experimental manipulation of feedback variability&#x2013;not for a role of this variability regulation mechanism in subsequent auditory-motor learning.</p>
</abstract>
<kwd-group>
<kwd>speech motor control</kwd>
<kwd>variability</kwd>
<kwd>adaptation</kwd>
<kwd>auditory feedback</kwd>
<kwd>acoustics</kwd>
<kwd>articulation</kwd>
</kwd-group>
<contract-num rid="cn001">R01DC014510</contract-num>
<contract-sponsor id="cn001">National Institutes of Health<named-content content-type="fundref-id">10.13039/100000002</named-content></contract-sponsor>
<counts>
<fig-count count="9"/>
<table-count count="1"/>
<equation-count count="0"/>
<ref-count count="56"/>
<page-count count="17"/>
<word-count count="11976"/>
</counts>
</article-meta>
</front>
<body>
<sec id="S1" sec-type="intro">
<title>Introduction</title>
<p>Over the years, the variability involved in human speech production has generated substantial empirical and theoretical interest. Both the physiological processes and acoustic output of speech production are inherently variable: even for a single speaker, no two repetitions of the same syllable are exactly the same in terms of muscle activation, kinematics, or acoustics (<xref ref-type="bibr" rid="B28">MacNeilage, 1970</xref>; <xref ref-type="bibr" rid="B37">Perkell and Klatt, 1986</xref>; <xref ref-type="bibr" rid="B26">Lindblom, 1990</xref>; <xref ref-type="bibr" rid="B35">Patri et al., 2015</xref>). Recently, it has started to become clear that such intra-individual variability at the behavioral level may reflect not only system noise but also functionally relevant adjustments in movement planning. Identifying the contribution of both these components will be critical for a better understanding of the sensorimotor control principles involved in spoken language.</p>
<p>To date, most experimental studies on the role of variability in speech production have taken an <italic>observational</italic> approach. That is, researchers typically have observed specific aspects of production variability in selected experimental conditions (without directly manipulating variability itself), and assessed the relationship with other measures of production or perception. For example, in the area of phonation, when subjects were asked to match a target tone by vocalizing with the same pitch and duration, those with greater production variability during the baseline phase exhibited stronger compensatory responses when unpredictable pitch perturbations were introduced in the auditory feedback signal (<xref ref-type="bibr" rid="B42">Scheerer and Jones, 2012</xref>). As an example from speech articulation, production variability for vowels has been shown to be linked to the speaker&#x2019;s categorical perceptual boundary between vowels (<xref ref-type="bibr" rid="B7">Chao et al., 2019</xref>). Various studies also examined production variability in relation to aspects of perception, but quantified variability <italic>across</italic> different speaking conditions or consonant contexts (e.g., how different is /&#x03B5;/ in &#x201C;bed&#x201D; vs. in &#x201C;tech&#x201D;), and thus did not address pure trial-to-trial variability in one particular phonetic context (e.g., <xref ref-type="bibr" rid="B38">Perkell et al., 2008</xref>; <xref ref-type="bibr" rid="B17">Franken et al., 2017</xref>).</p>
<p>Other groups have examined the potential relationship between observed trial-to-trial variability in speech acoustics and the extent of auditory-motor <italic>learning</italic> in a formant-shift adaptation paradigm. <xref ref-type="bibr" rid="B39">Purcell and Munhall (2006)</xref> reported a significant correlation between the lag 1 autocorrelation of trial-to-trial differences in a speaker&#x2019;s first formant (F1) during a baseline phase with unaltered auditory feedback and the extent of subsequent adaptation in response to a F1 perturbation. The relevance of this report is unclear, however, as calculating the lag 1 autocorrelation based on <italic>differences</italic> between neighboring trials can be a form of overdifferencing (<xref ref-type="bibr" rid="B13">Cryer and Chan, 2008</xref>). For example, it can be mathematically demonstrated that, after differencing, even a white noise time series has a lag 1 autocorrelation of &#x2013;0.5. Thus, finding a negative lag 1 autocorrelation based on differenced data does not necessarily mean that, in the original time series of formant data, trials were actually adjusted based on the preceding trial. In subsequent work, the same group quantified variability of vowel production as the standard deviations of a speaker&#x2019;s F1 and F2 distributions during the baseline phase (<xref ref-type="bibr" rid="B27">MacDonald et al., 2011</xref>). Using pooled data from seven experiments with a total of 116 participants, they found no significant correlation between these different metrics of variability and the extent of adaptation to an F1 perturbation. In a more recent study, the same group did report a significant correlation between baseline F1 standard deviation and F1 adaptation, but they also cautioned&#x2013;on the basis of a permutation test applied to the prior data&#x2013;that this was most likely a chance result (<xref ref-type="bibr" rid="B31">Nault and Munhall, 2020</xref>).</p>
<p>Thus, the question whether individual speakers&#x2019; baseline formant variability relates to their extent of auditory-motor learning in a formant-shift adaptation task remains unanswered to date. Interestingly, a study on upper limb sensorimotor control has suggested that reach movement trial-to-trial variability during a baseline phase does, in fact, facilitate early learning when adapting to a perturbing force field, possibly because greater variability offers more exploration of the task space (<xref ref-type="bibr" rid="B56">Wu et al., 2014</xref>). Even for upper limb movements, however, the generalizability and interpretation of this single study remain unclear (<xref ref-type="bibr" rid="B20">He et al., 2016</xref>; <xref ref-type="bibr" rid="B44">Singh et al., 2016</xref>; <xref ref-type="bibr" rid="B30">Murillo et al., 2017</xref>; <xref ref-type="bibr" rid="B46">Sternad, 2018</xref>; <xref ref-type="bibr" rid="B52">van der Vliet et al., 2018</xref>).</p>
<p>A more powerful approach toward addressing the issue of a potential relationship between sensorimotor variability and sensorimotor learning may consist of investigating variability with <italic>experimental</italic>, rather than observational, research methods. Direct experimental manipulation of inter-trial motor and/or sensory variability would allow one to ask multiple more specific questions. First, is inter-trial variability itself under active control by the central nervous system? In other words, can we find evidence of adjustments that compensate for increases or decreases in perceived variability of a specific performance measure? Second, does either the change in perceived variability of a performance measure or any active motor compensation for that perceived change affect sensorimotor adaptation in a new environment where that same aspect of performance is predictably perturbed?</p>
<p>To start investigating speech variability with such experimental methods, it is possible to adapt an approach taken in upper limb studies that magnified or attenuated visual feedback errors by a certain ratio (<xref ref-type="bibr" rid="B49">van Beers, 2009</xref>; <xref ref-type="bibr" rid="B55">Wong et al., 2009</xref>; <xref ref-type="bibr" rid="B36">Patton et al., 2013</xref>; <xref ref-type="bibr" rid="B51">van der Kooij et al., 2015</xref>). By aiming to manipulate the magnitude of feedback error in each trial, those studies also magnified and/or attenuated the <italic>dispersion</italic> of feedback error across trials. Hence, similar manipulations can be used to answer the above formulated question whether the inter-trial variability for a particular parameter of motor performance is actively controlled by the central nervous system. Specifically, motor behavior can be analyzed for any evidence of adjustments that compensate for the magnified or attenuated feedback variability. It should be noted that, in this context, a study&#x2019;s ability to both magnify and attenuate variability is critical from a methodological perspective. If an experimental paradigm only magnifies perceived variability by increasing the size of perceived movement errors, it is not possible to unambiguously attribute any resulting decrease in motor variability to the across-trials statistics <italic>per se</italic> vs. a preference for avoiding larger errors. If, on the other hand, a manipulation that attenuates perceived variability by minimizing perceived error leads to a compensatory <italic>increase</italic> in motor variability, then an interpretation based on the feedback statistics across trials is much more compelling as there are no theoretical reasons to expect a preference for avoiding smaller motor errors.</p>
<p>A reaching movement study by <xref ref-type="bibr" rid="B49">van Beers (2009)</xref> implemented such separate feedback conditions: movement endpoint errors were unaltered, reduced in magnitude by 50%, or increased in magnitude by 50%. Although the study did not specifically focus on compensation in terms of motor variability, <xref ref-type="bibr" rid="B49">van Beers (2009)</xref> found that the temporal structure of motor adjustments across trials differed among the visual feedback conditions: the sample lag 1 autocorrelation for movement endpoints was close to zero when errors in the feedback (and thus inter-trial variability) were not manipulated, negative when feedback errors were magnified, and positive when feedback errors were attenuated. The findings were interpreted in terms of which model of motor learning best explains subjects&#x2019; trial-to-trial adjustments, taking into account separate sources of central motor planning noise and peripheral motor execution noise. For natural movements with unperturbed feedback, <xref ref-type="bibr" rid="B49">van Beers (2009)</xref> concluded that trial-to-trial corrections are proportional to the magnitude of the previous error in such a way that movement variability is <italic>minimized</italic>, and it was suggested that this strategy is likely to underlie other forms of motor learning.</p>
<p>Lastly, a few upper limb studies have examined the effect of error magnification or attenuation on sensorimotor learning of a separate perturbation such as a visuomotor rotation. Results from those studies indicate that error magnification leads to more complete and faster adaptation whereas error attenuation has the opposite effect (<xref ref-type="bibr" rid="B36">Patton et al., 2013</xref>; <xref ref-type="bibr" rid="B51">van der Kooij et al., 2015</xref>). Despite this observed difference in adaptation, it has been argued that the adaptive learning mechanism itself, as quantified by a simple state-space model with the two parameters retention rate and error sensitivity, would remained unchanged between the different sensory feedback conditions (<xref ref-type="bibr" rid="B51">van der Kooij et al., 2015</xref>). However, other models of sensorimotor learning suggest that important parameters such as error sensitivity may be influenced by the prior history of feedback errors, a mechanism not captured by the simple state-space model (<xref ref-type="bibr" rid="B21">Herzfeld et al., 2014</xref>). Clearly, the effect of experimental manipulations of error magnitude and inter-trial variability on sensorimotor learning remains poorly understood even for upper limb movements.</p>
<p>Unfortunately, for speech articulation, work with experimental manipulations of feedback variability is only just starting to appear (see <xref ref-type="bibr" rid="B47">Tang et al., 2021</xref>), and the effect of manipulating the inter-trial variability of a specific parameter (e.g., frequency of one or more formants) on sensorimotor adaptation to a separate, predictable perturbation of the same parameter (e.g., a consistent formant shift) remains entirely unexplored. We therefore investigated whether an experimental magnification or attenuation of <italic>perceived</italic> inter-trial formant variability during speech production (a) leads to compensatory adjustments in <italic>produced</italic> formant variability, (b) induces changes in the temporal structure of formant adjustments across productions, and (c) affects subsequent auditory-motor learning when the speaker is exposed to a predictable formant-shift perturbation. Here, as the first step in this line of work, we implemented a relatively short-term formant variability manipulation (75 trials) and we looked for an effect on formant-shift adaptation in a subsequent task.</p>
</sec>
<sec id="S2" sec-type="materials|methods">
<title>Materials and Methods</title>
<sec id="S2.SS1">
<title>General Procedure</title>
<p>Twenty-eight right-handed adult native speakers of American English (20 women, 8 men, age <italic>M</italic> = 22.93 years, <italic>SD</italic> = 3.93years, range = 18&#x2013;31) with no self-reported history of speech, hearing or neurological disorders participated after providing written informed consent (all procedures were approved by the Institutional Review Board at the University of Washington). Based on a pure tone hearing screening, all participants had monaural thresholds at or below 20 dB HL at all octave frequencies from 250 Hz to 4 kHz in both ears.</p>
<p>The experiment was conducted in a sound-attenuated booth. First, participants completed a practice session with unaltered auditory feedback to familiarize themselves with the instrumentation set-up by producing 7 blocks of three target words. Each block consisted of the monosyllabic words &#x201C;talk,&#x201D; &#x201C;tech,&#x201D; and &#x201C;tuck&#x201D; in randomized order. These words were presented individually on a monitor in front of the participant, each word remaining visible for 3 s. To help participants maintain a consistent speaking style, visual feedback about speech intensity and duration was presented on the monitor after each production. The target intensity was between 72 and 80 dB SPL, and the target vowel duration was 100&#x2013;400 ms.</p>
<p>The actual experiment then included a <italic>Pre-test</italic> and two versions of a <italic>Variability</italic> task that were each immediately followed by an <italic>Adaptation</italic> task (<xref ref-type="fig" rid="F1">Figure 1A</xref>). The <italic>Pre-test</italic> served to determine each participant&#x2019;s median frequencies for the first and second formant (F1, F2) for the three target words (details below). During productions of the same words in the <italic>Variability</italic> tasks, inter-trial formant variability in the auditory feedback was either manipulated (<italic>magnified</italic> for one group of 4 men and 10 women, <italic>attenuated</italic> for the other group of 4 men and 10 women) or left unaltered (a control condition completed by both groups). Each <italic>Variability</italic> task was followed by an <italic>Adaptation</italic> task during which participants again produced the same target words but this time while hearing auditory feedback with a consistent upward perturbation of F1 and F2 (details below). The order of completing the manipulated and control versions of the <italic>Variability</italic> task (each immediately followed by an identical Adaptation task) was counterbalanced across participants.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption><p><bold>(A)</bold> Speech tasks completed by two groups of participants. Within each group, order of the experimental condition (Magnified or Attenuated feedback variability) and the Control condition was counterbalanced across participants. <bold>(B)</bold> Instrumentation setup. <bold>(C)</bold> Example spectrograms of Difference-shifted trials in the Magnified and Attenuated conditions of the <italic>Variability</italic> task. Dashed yellow line: pre-test median formant frequencies (F1 and F2, in Hz). Solid magenta and orange lines: produced (left of arrow) and heard (right of arrow) formants in the Magnified and Attenuated conditions. <bold>(D)</bold> Time course of the formant-shift feedback perturbation in the <italic>Adaptation</italic> task.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-16-890065-g001.tif"/>
</fig>
<p>In all of the above tasks, each participant&#x2019;s speech output was captured with a microphone (SM 58, Shure) positioned 15 cm from the mouth and connected to an audio interface (Babyface Pro, RME, Haimhausen, Germany) and computer located outside the soundbooth (<xref ref-type="fig" rid="F1">Figure 1B</xref>). The computer used MATLAB (The MathWorks, Natick, MA, United States) to present the visual stimuli, manipulate real-time auditory feedback when necessary, and record the participant&#x2019;s speech. Auditory feedback manipulations were implemented with the publicly available MATLAB software &#x2018;&#x2018;Audapter&#x2019;&#x2019;<sup><xref ref-type="fn" rid="footnote1">1</xref></sup> (<xref ref-type="bibr" rid="B5">Cai et al., 2008</xref>; <xref ref-type="bibr" rid="B48">Tourville et al., 2013</xref>). The output of the audio interface was amplified (HeadAmp6 Pro, ART ProAudio, Niagara Falls, NY, United States), and played back to the participant <italic>via</italic> insert earphones (ER-3A, Etymotic Research Inc., Grove Village, IL, United States). Before each participant&#x2019;s experiment, the feedback system was calibrated such that speech input with an intensity of 75 dB SPL at the microphone resulted in 72 dB SPL output in the earphones (<xref ref-type="bibr" rid="B12">Cornelisse et al., 1991</xref>). For this calibration procedure, the intensity of the auditory feedback in the earphones was measured using a 2 cc coupler (Type 4946, Bruel &#x0026; Kjaer Inc., Norcross, GA, United States) connected to a sound level meter (Type 2250A Hand Held Analyzer with Type 4947 &#x00BD;&#x2033; Pressure Field Microphone, Bruel &#x0026; Kjaer Inc., Norcross, GA, United States).</p>
<sec id="S2.SS1.SSS1">
<title>Pre-test</title>
<p>In the <italic>Pre-test</italic>, participants produced 30 blocks of the three target words with unaltered auditory feedback. During the production of each word, F1 and F2 were tracked by Audapter in real time. After the task was completed, a custom-written MATLAB script extracted the average F1 and F2 values (in Hz) across the middle portion of each production (defined as the window 40&#x2013;60% into the vowel), calculated the across-trials median F1 and F2 for each of the participant&#x2019;s vowels /&#x0254;/ (&#x201C;talk&#x201D;), /&#x03B5;/ (&#x201C;tech&#x201D;) and /&#x2227;/ (&#x201C;tuck&#x201D;), and identified the actual production closest to the pair of F1 and F2 median values for each vowel (closeness was defined based on Euclidean distance in F1-F2 space). The mid-vowel F1 and F2 values from the participant&#x2019;s three productions identified in this manner&#x2013;productions hereafter referred to as the pre-test medians for each vowel&#x2013;were used to determine the magnitude of the feedback variability manipulation in the <italic>Variability</italic> task. There was a short break (&#x223C;2 min) between the <italic>Pre-test</italic> and the first <italic>Variability</italic> task.</p>
</sec>
<sec id="S2.SS1.SSS2">
<title>Variability Task</title>
<p>Participants performed the <italic>Variability</italic> task once with auditory feedback in which F1 and F2 variability was experimentally manipulated (either <italic>magnified</italic> or <italic>attenuated</italic>, depending on the participant&#x2019;s group assignment) and once with unaltered auditory feedback as a control condition. In each <italic>Variability</italic> task, they produced 25 blocks of the three target words (for this first study with a variability perturbation, the number of trials was chosen based on published data regarding the number of trials that is sufficient for participants to reach maximum compensation in studies with other perturbations; see, for example, <xref ref-type="bibr" rid="B24">Kim et al., 2020a</xref>). In the <italic>magnified</italic> and <italic>attenuated</italic> conditions, formant variability in the auditory feedback was manipulated by modifying the difference between the formants in a given trial and the pre-test median for that vowel.</p>
<p>Specifically, a new mode of formant shifting, Difference-shift, was implemented by modifying Audapter&#x2019;s source code. In the new Difference-shift mode, the user supplies a target frequency for each formant (<italic>F<sup>T</sup></italic>) and a modification ratio (&#x03C1;). Within each frame, Audapter shifts the formant frequencies according to the equation <italic>F<sup>fb</sup></italic> = <italic>F<sup>T</sup></italic>+&#x03C1;&#x00D7;(<italic>F<sup>c</sup></italic>&#x2212;<italic>F<sup>T</sup></italic>),where <italic>F<sup>fb</sup></italic> is the formant frequency in the feedback and <italic>F<sup>c</sup></italic> is the formant frequency of the current production (both in Hz). Thus, Difference-shift modifies the difference between the current formant value and the target frequency by the modification ratio. For example, if the user enters 550 Hz as the target frequency for F1 and &#x03C1; = 2.5, then for an actual F1 value of 600 Hz, the Difference-shift mode shifts the output F1 to 675 Hz (550 + 2.5 50). When &#x03C1;=1, the Difference-shift mode magnifies the difference between the produced formant value and the target frequency, whereas the difference is attenuated when &#x03C1;&#x003C;1.</p>
<p>In both the Magnified and the Attenuated conditions, the pre-test median of F1 and F2 for each vowel was supplied as the target formant frequency <italic>F<sup>T</sup></italic>. To magnify the difference between the current production and the target formant frequency, &#x03C1; was set to 2.5 in the Magnified condition. To minimize the difference between the current production and the target, &#x03C1; was set to 0 in the Attenuated condition. Examples of individual productions and the corresponding manipulated feedback for each condition are included in <xref ref-type="fig" rid="F1">Figure 1C</xref>. Note that if &#x03C1;=0, the Difference-shift would theoretically always shift the formant frequency to the target frequency, regardless of the current production. However, due to the intrinsic limitations of real-time formant tracking and the digital filtering techniques used to alter the signal, the actual ratio between produced frequency and Difference-shifted output frequency is not always identical to the supplied modification ratio. Given this situation that, in reality, &#x03C1;=0 reduces (but does not completely eliminate) feedback variability, it was chosen as the preferred ratio for the Attenuated condition. The overall effectiveness of the feedback perturbation for magnifying and attenuating feedback formant variability is described below in the Section &#x201C;Results.&#x201D;</p>
</sec>
<sec id="S2.SS1.SSS3">
<title>Adaptation Task</title>
<p>Each <italic>Adaptation</italic> task followed immediately after one of the <italic>Variability</italic> tasks, and was identical after the manipulated and control versions of the <italic>Variability</italic> task. In both cases, it consisted of a perturbation phase (25 blocks) and an after-effect phase (15 blocks) (<xref ref-type="fig" rid="F1">Figure 1D</xref>). No variability manipulation was applied, but, at the start of the perturbation phase, a sudden 250 cents<sup><xref ref-type="fn" rid="footnote2">2</xref></sup> upshift of F1 and F2 was introduced by Audapter. This formant shift was turned off, and participants received unaltered auditory feedback, during the after-effect phase. There was a short break (&#x223C;2 min) between the end of the first <italic>Adaptation</italic> task and the beginning of the second <italic>Variability</italic> task.</p>
</sec>
</sec>
<sec id="S2.SS2">
<title>Data Extraction and Analysis</title>
<p>The speech signal from all tasks (<italic>Pre-test</italic> task, <italic>Variability</italic> tasks, and <italic>Adaptation</italic> tasks) was digitized by Audapter. Using a custom-written MATLAB script, we examined the production data from all tasks offline to exclude productions containing production errors (e.g., mispronunciations or yawning; 0.45% of productions were rejected for this reason), manually marked the onset and offset of the vowel in each production based on visual inspection of its waveform and spectrogram, and extracted the first two formant frequencies (F1 and F2) as tracked by the linear predictive coding algorithm implemented in Praat (<xref ref-type="bibr" rid="B3">Boersma, 2001</xref>). To disentangle feedforward adaptive learning vs. online feedback-driven corrections within trials, F1 and F2 formant values for each trial were extracted both across an initial portion of the vowel (5&#x2013;30% into its total duration) and a middle portion of the vowel (40&#x2013;60% into total duration). Additionally, to verify accuracy of the auditory feedback manipulation in the experimental conditions of the <italic>Variability</italic> task (Magnified and Attenuated variability), we extracted F1 and F2 also across the same middle portion of the vowel in the recorded feedback signal.</p>
<p>Statistical analyses for the <italic>Variability</italic> task and the <italic>Adaptation</italic> task made use of paired two-sample <italic>t</italic>-tests or, in a few cases, one-sample <italic>t</italic>-tests, with the significance level set at 0.05. When multiple statistical comparisons were carried out as one family of tests, <italic>p</italic>-values were adjusted with the Holm&#x2013;Bonferroni method (<xref ref-type="bibr" rid="B22">Holm, 1979</xref>). Cohen&#x2019;s <italic>d</italic> was used for effect size calculations (<xref ref-type="bibr" rid="B11">Cohen, 1988</xref>). All statistical tests were conducted in the R software (<xref ref-type="bibr" rid="B40">R Core Team, 2019</xref>).</p>
<sec id="S2.SS2.SSS1">
<title>Analysis of the Variability Task</title>
<p>Formant frequencies measured for the initial and middle portions of vowels from the <italic>Variability</italic> task were normalized by conversion from Hz to cents. The medians (F1 and F2) of each vowel from each subject&#x2019;s pre-test productions, also measured offline across the initial and the middle portions separately, were chosen as the reference frequency for the conversion. Similarly, the formants measured from the middle portion of the vowel in the auditory feedback signal were also converted with reference to each subject&#x2019;s pre-test median frequencies for the middle portion.</p>
<p>A primary focus of analysis for the <italic>Variability</italic> task was the participants&#x2019; production variability. To quantify this production variability with a measure directly related to the nature of the perturbation itself (i.e., distance to the pre-test median formants), we formulated a distance index (DI), <inline-formula><mml:math id="INEQ15"><mml:mrow><mml:mrow><mml:mi>D</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mpadded width="+3.3pt"><mml:mi>I</mml:mi></mml:mpadded></mml:mrow><mml:mo rspace="10.8pt">=</mml:mo><mml:msqrt><mml:mrow><mml:mrow><mml:mi>F</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mpadded width="+3.3pt"><mml:msup><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:msup></mml:mpadded></mml:mrow><mml:mo rspace="5.8pt">+</mml:mo><mml:mrow><mml:mi>F</mml:mi><mml:mo>&#x2062;</mml:mo><mml:msup><mml:mn>2</mml:mn><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mrow></mml:msqrt></mml:mrow></mml:math></inline-formula>, where F1 and F2 are a trial&#x2019;s formant frequencies already expressed in cents relative to the pre-test median. For each production, two DI&#x2019;s, DI<sub>initial</sub> and DI<sub>mid</sub>, were calculated with the formant values that had been extracted from the non-overlapping initial and middle portions of that trial&#x2019;s vowel. For the auditory feedback signal, there was only one DI measurement per trial, DI<sub>fb</sub>, as formant frequencies had been extracted only for the middle portion of the vowel.</p>
<p>First, to verify the effectiveness of our formant feedback variability magnification and attenuation by the Difference-shift implementation in Audapter, the ratio between the average DI<sub>fb</sub> and average DI<sub>mid</sub> of each participant&#x2019;s experimental <italic>Variability</italic> task was compared to the ideal ratio based on the perturbation algorithm (assuming perfect formant tracking and signal processing). Second, to examine the effect of feedback variability manipulation on production variability (<xref ref-type="bibr" rid="B55">Wong et al., 2009</xref>), we compared both DI<sub>initial</sub> and DI<sub>mid</sub> between the Control condition and the experimental (Magnified or Attenuated) conditions. To explore the possibility of gradual changes in production variability during the course of the <italic>Variability</italic> task, these variables were considered not only for the whole task (25 blocks of 3 trials each) but also block-by-block and stage-by-stage (with a stage operationally defined as a series of 5 consecutive blocks). Third, to examine possible online feedback-based corrections in response to the variability manipulations, we also calculated the within-trial difference between DI<sub>initial</sub> and DI<sub>mid</sub> [note that this approach shows similarities with the &#x201C;centering&#x201D; measure used in previous studies of online feedback corrections (<xref ref-type="bibr" rid="B33">Niziolek et al., 2013</xref>; <xref ref-type="bibr" rid="B32">Niziolek and Kiran, 2018</xref>), but differs from it in that our DI measures determine each trial&#x2019;s distance to the median production from the <italic>Pre-test</italic> in cents rather than distance to the median production of the analyzed dataset itself in mels]. For each experimental condition (Magnified, Attenuated) and each control condition (completed by the Magnified and Attenuated groups separately), we used one-sample <italic>t</italic>-tests to determine whether the within trial changes in DI were statistically significantly different from zero (i.e., whether or not &#x201C;centering&#x201D; toward the pre-test median occurred). For each group separately, we then used paired <italic>t</italic>-tests to determine whether any within-trial changes differed between the experimental and control condition.</p>
<p>Although analogous to the nature of the variability perturbation itself, one potential problem with the DI-based analysis is that it is theoretically possible for a participant to increase or decrease the average distance between their trial formant frequencies and the pre-test medians without increasing the actual dispersion of these trials in two-dimensional (F1, F2) acoustic space. For example, although extremely unlikely for real speech, it is theoretically possible that the formants for all trials could be moved further away from the pre-test median (thereby increasing DI) but always to the same location in acoustic space. For this reason, we followed up on statistically significant DI effects by also determining for each participant the size of the area in acoustic space covered by the relevant productions (i.e., trials produced in the Control condition or in a given stage of the experimental conditions). The size of this area was determined by means of 95% confidence ellipses, calculated based on formant frequencies from the initial portion of the vowels.</p>
<p>A secondary focus of the <italic>Variability</italic> task was to investigate possible effects of the variability manipulations on the temporal structure of formant adjustments across trials. Consistent with the approach used in previous non-speech studies (<xref ref-type="bibr" rid="B49">van Beers, 2009</xref>; <xref ref-type="bibr" rid="B52">van der Vliet et al., 2018</xref>), we compared between Control and experimental conditions the sample lag 1 autocorrelation function, ACF(1), calculated for the sequence of averaged formant frequencies (i.e., mean of F1 and F2) obtained at the initial portion of the vowel in each trial. Formally, <inline-formula><mml:math id="INEQ16"><mml:mrow><mml:mrow><mml:mi>A</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>C</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>F</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>1</mml:mn><mml:mo rspace="5.8pt">)</mml:mo></mml:mrow></mml:mrow><mml:mo rspace="5.8pt">=</mml:mo><mml:mrow><mml:mfrac><mml:mn>1</mml:mn><mml:mi>N</mml:mi></mml:mfrac><mml:mo>&#x2062;</mml:mo><mml:mfrac><mml:mrow><mml:mstyle displaystyle="false"><mml:msubsup><mml:mo largeop="true" symmetric="true">&#x2211;</mml:mo><mml:mrow><mml:mpadded width="+3.3pt"><mml:mi>n</mml:mi></mml:mpadded><mml:mo rspace="5.8pt">=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mstyle><mml:msup><mml:mi/><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mi>F</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow></mml:mrow><mml:mo>-</mml:mo><mml:mover accent="true"><mml:mi>F</mml:mi><mml:mo>&#x00AF;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mi>F</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mi>n</mml:mi><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo>-</mml:mo><mml:mover accent="true"><mml:mi>F</mml:mi><mml:mo>&#x00AF;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mstyle displaystyle="false"><mml:msubsup><mml:mo largeop="true" symmetric="true">&#x2211;</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>N</mml:mi></mml:msubsup></mml:mstyle><mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mi>F</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mi>n</mml:mi><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo>-</mml:mo><mml:mover accent="true"><mml:mi>F</mml:mi><mml:mo>&#x00AF;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac></mml:mrow></mml:mrow></mml:math></inline-formula>, where <italic>N</italic> = 75, <inline-formula><mml:math id="INEQ17"><mml:mrow><mml:mi>F</mml:mi><mml:mo stretchy='false'>[</mml:mo><mml:mi>n</mml:mi><mml:mo stretchy='false'>]</mml:mo><mml:mo>=</mml:mo><mml:mo>&#x2009;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>F</mml:mi><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msub><mml:mo stretchy='false'>[</mml:mo><mml:mi>n</mml:mi><mml:mo stretchy='false'>]</mml:mo><mml:mo>&#x2009;</mml:mo><mml:mo>+</mml:mo><mml:mo>&#x2009;</mml:mo><mml:mi>F</mml:mi><mml:msub><mml:mn>2</mml:mn><mml:mrow><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msub><mml:mo stretchy='false'>[</mml:mo><mml:mi>n</mml:mi><mml:mo stretchy='false'>]</mml:mo><mml:mo stretchy='false'>)</mml:mo><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:math></inline-formula>, and <inline-formula><mml:math id="INEQ18"><mml:mrow><mml:mpadded width="+3.3pt"><mml:mover accent="true"><mml:mi>F</mml:mi><mml:mo>&#x00AF;</mml:mo></mml:mover></mml:mpadded><mml:mo rspace="5.8pt">=</mml:mo><mml:mrow><mml:mfrac><mml:mn>1</mml:mn><mml:mi>N</mml:mi></mml:mfrac><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:msubsup><mml:mo largeop="true" symmetric="true">&#x2211;</mml:mo><mml:mrow><mml:mpadded width="+3.3pt"><mml:mi>n</mml:mi></mml:mpadded><mml:mo rspace="5.8pt">=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>N</mml:mi></mml:msubsup><mml:mrow><mml:mi>F</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mi>n</mml:mi><mml:mo stretchy="false">]</mml:mo></mml:mrow></mml:mrow></mml:mrow></mml:mrow></mml:mrow></mml:math></inline-formula>.</p>
</sec>
<sec id="S2.SS2.SSS2">
<title>Analysis of the Adaptation Task</title>
<p>Given that adaptation refers to adjustments in movement planning based on prior experience (as opposed to online feedback-driven corrections), only the formant frequencies measured at the initial portion of the vowel were used for analysis of the <italic>Adaptation</italic> task. These formant frequencies were normalized to cents with reference to the median formants of each vowel in blocks 16&#x2013;25 of the <italic>Variability</italic> task immediately prior to the onset of the <italic>Adaptation</italic> task. The frequencies of F1 and F2, in cents, were averaged for each trial as in several of our prior studies (e.g., <xref ref-type="bibr" rid="B24">Kim et al., 2020a</xref>; <xref ref-type="bibr" rid="B43">Shiller et al., 2020</xref>).</p>
<p>We compared three metrics between the perturbation phases from the Control and experimental conditions: early adaptation extent, early adaptation rate, and final adaptation extent. Early adaptation extent was calculated by determining the average formant frequency across the first 15 trials of the perturbation phase. Early adaptation rate was defined as the slope of a linear regression function based on the formant frequencies of the same 15 trials. Final adaptation extent was calculated by determining the average formant frequency across the last 15 trials of the perturbation phase of the task.</p>
</sec>
</sec>
</sec>
<sec id="S3" sec-type="results">
<title>Results</title>
<sec id="S3.SS1">
<title>Variability Task</title>
<sec id="S3.SS1.SSS1">
<title>Effectiveness of the Feedback Variability Manipulations</title>
<p>Individual participant data for DI calculated for both the produced and heard trials from the <italic>Variability</italic> task are presented in <xref ref-type="fig" rid="F2">Figure 2</xref>. <xref ref-type="fig" rid="F2">Figures 2A,C</xref> each show that the feedback manipulation was effective for the two selected participants from the Magnified and Attenuated conditions, respectively. <xref ref-type="fig" rid="F2">Figures 2B,D</xref> show for <italic>all</italic> individual participants the ratio between the average DI of the formants in the manipulated feedback, DI<sub>fb</sub>, and that of the produced formants, DI<sub>mid</sub>, for the Magnified and Attenuated conditions, respectively. In the Magnified condition, the group mean of this ratio was 2.52 (<italic>SD</italic> = 0.27), a value very close to the intended modification ratio &#x03C1;=2.5 (which is also the theoretical value of DI<sub>fb</sub>/DI<sub>mid</sub> if the Difference-shift had worked perfectly in every frame of every trial). In the Attenuated condition (ratio &#x03C1;=0), however, there was one outlier participant (DI<sub>fb</sub>/DI<sub>prod</sub> = 1.202) for whom the Difference-shift mode failed to achieve the goal of attenuating formant variability in the auditory feedback. With the outlier removed, the group mean of the DI<sub>fb</sub>/DI<sub>mid</sub> ratio was 0.49 (<italic>SD</italic> = 0.14) and all remaining ratios were less than 1, confirming that the goal of attenuating feedback variability was achieved. The data from the participant with the unsuccessful feedback perturbation in the Attenuated condition were excluded from all further analyses.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption><p><bold>(A,C)</bold> Example individual participant data for production and feedback Distance Index (DI) of each trial in the <italic>Variability</italic> task under Magnified and Attenuated conditions. <bold>(B,D)</bold> Boxplots with symbols depicting each participant&#x2019;s ratio between average feedback DI and average production DI (both measured mid-vowel) across all trials of the <italic>Variability</italic> task with Magnified or Attenuated feedback variability.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-16-890065-g002.tif"/>
</fig>
</sec>
<sec id="S3.SS1.SSS2">
<title>Production Variability</title>
<p>The first set of production variability analyses compared the Control condition with both experimental conditions at the whole-task level for the target vowel&#x2019;s initial portion (DI<sub>initial</sub>, <xref ref-type="fig" rid="F3">Figure 3A</xref> for the Magnified condition, <xref ref-type="fig" rid="F3">Figure 3C</xref> for the Attenuated condition) and middle portion (DI<sub>mid</sub>, <xref ref-type="fig" rid="F3">Figures 3B,D</xref>). As compared with the Control condition, no statistically significant change in DI<sub>initial</sub> was found for either the Magnified or the Attenuated condition [<italic>t</italic>(13) = &#x2013;0.282, <italic>p</italic> = 0.782, <italic>d</italic> = 0.075, and <italic>t</italic>(12) = 1.358, <italic>p</italic> = 0.200, <italic>d</italic> = &#x2013;0.376, respectively]. Similarly, there was also no significant change in DI<sub>mid</sub> for either the Magnified or Attenuated condition [<italic>t</italic>(13) = 0.231, <italic>p</italic> = 0.821, <italic>d</italic> = &#x2013;0.062, and <italic>t</italic>(12) = 2.122, <italic>p</italic> = 0.055, <italic>d</italic> = 0.588, respectively].</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption><p><bold>(A,C)</bold> Boxplots with symbols depicting each participant&#x2019;s average DI<sub>initial</sub> for the entire <italic>Variability</italic> task in Control and Magnified or Control and Attenuated conditions. <bold>(B,D)</bold> Boxplots with each participant&#x2019;s average DI<sub>mid</sub> for the entire <italic>Variability</italic> task in the same conditions.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-16-890065-g003.tif"/>
</fig>
<p>The second set of production variability analyses examined whether a response to the auditory feedback manipulations might develop over time with continuing exposure. Therefore, these analyses considered the time course of the DI<sub>initial</sub> and DI<sub><italic>mid</italic></sub> variables per block of 3 trials and per stage of 5 blocks. <xref ref-type="fig" rid="F4">Figure 4</xref> shows group data for the change in DI<sub>initial</sub> from block to block (<xref ref-type="fig" rid="F4">Figures 4A,B</xref>) and stage to stage (<xref ref-type="fig" rid="F4">Figures 4C,D</xref>) under the Control and experimental conditions. For the group that completed Control and Magnified conditions, the data show no change in formant production DI<sub>initial</sub> within either of those conditions. Statistical comparisons of DI<sub>initial</sub> between the first stage and each of the following stages confirmed the absence of an adjustment in this distance metric with Magnified feedback variability (<xref ref-type="table" rid="T1">Table 1</xref> and <xref ref-type="fig" rid="F4">Figure 4C</xref>). In contrast, for the group that completed Control and Attenuated conditions, DI<sub>initial</sub> showed a statistically significant increase from Stage 1 to Stage 2 and from Stage 1 to Stage 3 in the condition with Attenuated feedback variability whereas no statistically significant change was observed in the same group&#x2019;s Control condition (<xref ref-type="table" rid="T1">Table 1</xref> and <xref ref-type="fig" rid="F4">Figure 4D</xref>). Visualizations of the Attenuated condition individual participant data for DI<sub>initial</sub> in Stage 1 and Stage 3, and of the extent and direction for individual changes in this variable over the same time period, are included in <xref ref-type="fig" rid="F4">Figures 4E&#x2013;H</xref> (analysis techniques based on <xref ref-type="bibr" rid="B54">Wilcox and Erceg-Hurn, 2012</xref>; <xref ref-type="bibr" rid="B2">Bieniek et al., 2016</xref>; <xref ref-type="bibr" rid="B41">Rousselet et al., 2017</xref>). The data show a robust trend across individuals as 11 of 13 participants increased their formant production DI<sub>initial</sub> in the first half of the task with Attenuated feedback variability.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption><p><bold>(A,B)</bold> Change in DI<sub>initial</sub> across the Variability task by block (i.e., 3 trials) for the Magnified and Attenuated feedback variability conditions. Dots represent the group mean DI per block. Shaded regions indicate standard error of the mean (SEM). Solid lines are loess smoothed fits (span = 0.6). <bold>(C,D)</bold> Change in DI<sub>initial</sub> across the Variability task by stage (i.e., 15 trials) for the Magnified and Attenuated feedback variability conditions. Error bars indicate SEM. Asterisks indicate adjusted p &#x003C; 0.05 (see <xref ref-type="table" rid="T1">Table 1</xref>). <bold>(E&#x2013;H)</bold> Individual participant data for the significant change from Stage 1 to Stage 3 in the Attenuated condition: <bold>(E)</bold> Stripchart of DI<sub>initial</sub> in Stage 1 and Stage 3. Horizontal lines indicate deciles; bold line is the median. <bold>(F)</bold> Stripchart with each participant&#x2019;s Stage 1 and Stage 3 data linked. <bold>(G)</bold> Scatterplot of Stage 1 by Stage 3 data. The diagonal line denotes no difference between stages. Participants in the upper left half increased DI<sub>initial</sub> in Stage 3. Dashed lines mark quartiles. <bold>(H)</bold> Stripchart of the difference in DI<sub>initial</sub> between Stage 3 and Stage 1. Horizontal lines indicate deciles; the bold line is the median; the dashed line is at zero (no difference between stages).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-16-890065-g004.tif"/>
</fig>
<table-wrap position="float" id="T1">
<label>TABLE 1</label>
<caption><p>Adjusted <italic>p</italic>-values (paired <italic>t</italic>-tests, Holm&#x2013;Bonferroni method) for comparisons of DI<sub>initial</sub> (top section) and DI<sub>mid</sub> (bottom section) between Stage 1 (first 5 blocks of 3 trials) and each subsequent stage (also 15 trials) in the Control, Magnified, and Attenuated feedback variability conditions (the two participant groups completing Magnified or Attenuated variability conditions each completed their own Control conditions, labeled Control M and Control A).</p></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center">Stage 1 vs. 2</td>
<td valign="top" align="center">Stage 1 vs. 3</td>
<td valign="top" align="center">Stage 1 vs. 4</td>
<td valign="top" align="center">Stage 1 vs. 5</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center" colspan="4"><bold>DI<sub>initial</sub></bold></td>
</tr>
<tr>
<td valign="top" align="left">Control M</td>
<td valign="top" align="center"><italic>t</italic>(13) = &#x2013;1.519, <italic>p</italic> = 0.612, <italic>d</italic> = &#x2013;0.406</td>
<td valign="top" align="center"><italic>t</italic>(13) = &#x2013;1.278, <italic>p</italic> = 0.612, <italic>d</italic> = &#x2013;0.342</td>
<td valign="top" align="center"><italic>t</italic>(13) = &#x2013;0.234, <italic>p</italic> = 0.819, <italic>d</italic> = &#x2013;0.062</td>
<td valign="top" align="center"><italic>t</italic>(13) = &#x2013;1.514, <italic>p</italic> = 0.612, <italic>d</italic> = &#x2013;0.405</td>
</tr>
<tr>
<td valign="top" align="left">Magnified</td>
<td valign="top" align="center"><italic>t</italic>(13) = &#x2013;0.506, <italic>p</italic> = 1.000, <italic>d</italic> = &#x2013;0.135</td>
<td valign="top" align="center"><italic>t</italic>(13) = &#x2013;0.680, <italic>p</italic> = 1.000, <italic>d</italic> = &#x2013;0.182</td>
<td valign="top" align="center"><italic>t</italic>(13) = &#x2013;0.111, <italic>p</italic> = 1.000, <italic>d</italic> = &#x2013;0.030</td>
<td valign="top" align="center"><italic>t</italic>(13) = -1.717, <italic>p</italic> = 0.440, <italic>d</italic> = &#x2013;0.459</td>
</tr>
<tr>
<td valign="top" align="left">Control A</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;0.041, <italic>p</italic> = 1.000, <italic>d</italic> = &#x2013;0.011</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;0.561, <italic>p</italic> = 1.000, <italic>d</italic> = &#x2013;0.156</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;1.153, <italic>p</italic> = 1.000, <italic>d</italic> = &#x2013;0.320</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;0.944, <italic>p</italic> = 1.000, <italic>d</italic> = &#x2013;0.262</td>
</tr>
<tr>
<td valign="top" align="left">Attenuated</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;2.800, <italic>p</italic> = 0.048<xref ref-type="table-fn" rid="t1fn1">&#x002A;</xref>, <italic>d</italic> = &#x2013;0.777</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;3.189, <italic>p</italic> = 0.031<xref ref-type="table-fn" rid="t1fn1">&#x002A;</xref>, <italic>d</italic> = &#x2013;0.884</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;2.330, <italic>p</italic> = 0.076, <italic>d</italic> = &#x2013;0.646</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;2.051, <italic>p</italic> = 0.076, <italic>d</italic> = &#x2013;0.569</td>
</tr>
<tr>
<td/>
<td valign="top" align="center" colspan="4"><bold>DI<sub>mid</sub></bold></td>
</tr>
<tr>
<td valign="top" align="left">Control M</td>
<td valign="top" align="center"><italic>t</italic>(13) = &#x2013;0.426, <italic>p</italic> = 1.000, <italic>d</italic> = &#x2013;0.114</td>
<td valign="top" align="center"><italic>t</italic>(13) = &#x2013;0.077, <italic>p</italic> = 1.000, <italic>d</italic> = &#x2013;0.021</td>
<td valign="top" align="center"><italic>t</italic>(13) = &#x2013;1.347, <italic>p</italic> = 0.804, <italic>d</italic> = &#x2013;0.360</td>
<td valign="top" align="center"><italic>t</italic>(13) = &#x2013;0.918, <italic>p</italic> = 1.000, <italic>d</italic> = &#x2013;0.245</td>
</tr>
<tr>
<td valign="top" align="left">Magnified</td>
<td valign="top" align="center"><italic>t</italic>(13) = 0.473, <italic>p</italic> = 1.000, <italic>d</italic> = 0.126</td>
<td valign="top" align="center"><italic>t</italic>(13) = 0.676, <italic>p</italic> = 1.000, <italic>d</italic> = 0.181</td>
<td valign="top" align="center"><italic>t</italic>(13) = 0.501, <italic>p</italic> = 1.000, <italic>d</italic> = 0.134</td>
<td valign="top" align="center"><italic>t</italic>(13) = &#x2013;2.526, <italic>p</italic> = 0.101, <italic>d</italic> = &#x2013;0.675</td>
</tr>
<tr>
<td valign="top" align="left">Control A</td>
<td valign="top" align="center"><italic>t</italic>(12) = 0.659, <italic>p</italic> = 1.000, <italic>d</italic> = 0.183</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;0.126, <italic>p</italic> = 1.000, <italic>d</italic> = &#x2013;0.035</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;1.133, <italic>p</italic> = 1.000, <italic>d</italic> = &#x2013;0.314</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;0.427, <italic>p</italic> = 1.000, <italic>d</italic> = &#x2013;0.118</td>
</tr>
<tr>
<td valign="top" align="left">Attenuated</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;3.039, <italic>p</italic> = 0.021<xref ref-type="table-fn" rid="t1fn1">&#x002A;</xref>, <italic>d</italic> = &#x2013;0.843</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;4.996, <italic>p</italic> = 0.001<xref ref-type="table-fn" rid="t1fn1">&#x002A;</xref>, <italic>d</italic> = &#x2013;1.386</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;1.929, <italic>p</italic> = 0.078, <italic>d</italic> = &#x2013;0.535</td>
<td valign="top" align="center"><italic>t</italic>(12) = &#x2013;3.406, <italic>p</italic> = 0.016<xref ref-type="table-fn" rid="t1fn1">&#x002A;</xref>, <italic>d</italic> = &#x2013;0.945</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="t1fn1"><p><italic>Statistically significant differences (&#x002A;) were found only for the Attenuated feedback variability manipulation, in particular for the comparisons Stage 1 vs. Stage 2 and Stage 1 vs. Stage 3.</italic></p></fn>
</table-wrap-foot>
</table-wrap>
<p>Similar results were obtained when considering DI<sub>mid</sub> from block to block (<xref ref-type="fig" rid="F5">Figures 5A,B</xref>) and stage to stage (<xref ref-type="fig" rid="F5">Figures 5C,D</xref>): DI<sub>mid</sub> showed no change in either group&#x2019;s Control condition, also no change in the Magnified condition, but a statistically significant increase from Stage 1 to Stages 2, 3, and 5 in the Attenuated condition (<xref ref-type="table" rid="T1">Table 1</xref> and <xref ref-type="fig" rid="F5">Figure 5D</xref>). The individual participant data for Stage 1 and Stage 3 in this condition with Attenuated feedback variability show a highly consistent increase in formant production DI<sub>mid</sub> during the first half of the task (<xref ref-type="fig" rid="F5">Figures 5E&#x2013;H</xref>).</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption><p><bold>(A,B)</bold> Change in DI<sub>mid</sub> across the Variability task by block (i.e., 3 trials) for the Magnified and Attenuated feedback variability conditions. Dots represent the group mean DI per block. Shaded regions indicate standard error of the mean (SEM). Solid lines are loess smoothed fits (span = 0.6). <bold>(C,D)</bold> Change in DI<sub>mid</sub> across the Variability task by stage (i.e., 15 trials) for the Magnified and Attenuated feedback variability conditions. Error bars indicate SEM. Asterisks indicate adjusted p &#x003C; 0.05 (see <xref ref-type="table" rid="T1">Table 1</xref>). <bold>(E&#x2013;H)</bold> Individual participant data for the significant change from Stage 1 to Stage 3 in the Attenuated condition: <bold>(E)</bold> Stripchart of DI<sub>mid</sub> in Stage 1 and Stage 3. Horizontal lines indicate deciles; bold line is the median. <bold>(F)</bold> Stripchart with each participant&#x2019;s Stage 1 and Stage 3 data linked. <bold>(G)</bold> Scatterplot of Stage 1 by Stage 3 data. The diagonal line denotes no difference between stages. Participants in the upper left half increased DI<sub>mid</sub> in Stage 3. Dashed lines mark quartiles. <bold>(H)</bold> Stripchart of the difference in DI<sub>mid</sub> between Stage 3 and Stage 1. Horizontal lines indicate deciles; the bold line is the median; the dashed line is at zero (no difference between stages).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-16-890065-g005.tif"/>
</fig>
<p>We examined the change from DI<sub>initial</sub> to DI<sub>mid</sub> as an indicator of potential within-trial corrections in the conditions with Magnified or Attenuated formant variability in the auditory feedback. For participants assigned to the Attenuated group, within-trial changes were not statistically significantly different from zero for either the experimental condition [<italic>t</italic>(12) = 0.164, <italic>p</italic> = 0.872, <italic>d</italic> = 0.046] or the control condition [<italic>t</italic>(12) = &#x2013;1.285, <italic>p</italic> = 0.446, <italic>d</italic> = &#x2013;0.356]. For participants in the Magnified group, within-trial changes were statistically significant, but this was the case for both the experimental condition [<italic>t</italic>(13) = &#x2013;4.117, <italic>p</italic> = 0.002, <italic>d</italic> = &#x2013;1.100] and the control condition [<italic>t</italic>(13) = &#x2013;3.650, <italic>p</italic> = 0.003, <italic>d</italic> = &#x2013;0.975]. For neither group were within-trial changes in the experimental condition statistically different from those in the control condition with unaltered feedback variability [Attenuated group: <italic>t</italic>(12) = &#x2013;0.997, <italic>p</italic> = 0.339, <italic>d</italic> = &#x2013;0.288; Magnified group: <italic>t</italic>(13) = &#x2013;0.962, <italic>p</italic> = 0.354, <italic>d</italic> = 0.264].</p>
<p>Given that the Attenuated condition showed a statistically significant increase in DI<sub>initial</sub> (as well as DI<sub>mid</sub>) from Stage 1 to Stage 3, <xref ref-type="fig" rid="F6">Figure 6</xref> shows the individual participants&#x2019; inter-trial dispersion of formant frequencies in 2D (F1, F2) acoustic space for Stages 1 and 3 of the Attenuated variability condition together with equivalent data from the <italic>Pre-test</italic>. All data were extracted from the initial portion of the vowels. Although the comparison of 95% confidence ellipse areas for Stage 1 versus Stage 3 did not reach statistical significance [<italic>t</italic>(12) = &#x2013;1.894, <italic>p</italic> = 0.083], this comparison was associated with a medium effect size (<italic>d</italic> = &#x2013;0.525), and 9 of 13 individual participants increased the ellipse area in Stage 3 as compared with Stage 1. Of the four participants who decreased ellipse size, only two showed a change that fell within the range of changes (but with opposite sign) observed for the subjects with increasing ellipses; the other two subjects showed only minimal changes.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption><p>Individual participant data (one participant per panel) for inter-trial formant dispersion in acoustic vowel space (F1 by F2). Data based on 95% confidence ellipses, calculated for formant frequencies extracted from the initial portion of the vowels. Each participant&#x2019;s data from Stages 1 and 3 (15 trials per stage) in the Attenuated feedback variability condition are shown together with their data from the <italic>Pre-test</italic> (90 trials). Nine of 13 participants increased ellipse area in Stage 3 as compared with Stage 1. Participants are ordered (by row) from greatest to smallest ellipse area increase.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-16-890065-g006.tif"/>
</fig>
</sec>
<sec id="S3.SS1.SSS3">
<title>Autocorrelation Structure</title>
<p>To assess the temporal structure of formant adjustments across the entire series of productions in the manipulated auditory feedback conditions, we determined the sample lag 1 autocorrelation [ACF(1)] of the time series consisting of averaged F1 and F2 values from the initial vowel portion of each trial in the <italic>Variability</italic> task (<xref ref-type="fig" rid="F7">Figure 7</xref>). It should be noted that the large sample 95% confidence interval of ACF(1) for a white noise process with sample size <italic>N</italic> = 75 (i.e., the number of trials in each analyzed time series) is (-0.22, 0.22) (<xref ref-type="bibr" rid="B4">Brockwell and Davis, 2016</xref>). Most of the individual ACF(1) data from all conditions in the current study fell within this bound, indicating that, from a statistical perspective, it is likely that most production sequences were generated by white noise processes. There were no statistically significant differences in ACF(1) between either of the two experimental conditions and the Control condition [Magnified: <italic>t</italic>(13) = 0.670, <italic>p</italic> = 0.515, <italic>d</italic> = 0.179; Attenuated: <italic>t</italic>(12) = &#x2013;0.324, <italic>p</italic> = 0.752, <italic>d</italic> = 0.090].</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption><p>Sample lag 1 autocorrelation functions [ACF(1)] for formant data measured in the initial portion of the vowel and averaged across F1 and F2 for Control versus Magnified <bold>(A)</bold> and Control versus Attenuated <bold>(B)</bold> conditions of the <italic>Variability</italic> task. Dashed lines indicate the large sample 95% confidence interval of ACF(1) for a white noise process with sample size 75 (the number of trials per condition). Each dot represents an individual participant.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-16-890065-g007.tif"/>
</fig>
</sec>
</sec>
<sec id="S3.SS2">
<title>Adaptation Task</title>
<p><xref ref-type="fig" rid="F8">Figure 8</xref> shows group mean formant frequencies produced throughout the <italic>Adaptation</italic> tasks that followed immediately after different conditions of the <italic>Variability</italic> task (data are in cents relative to the end of the preceding <italic>Variability</italic> task, measured at the initial portion of the vowel, and averaged across F1 and F2 and across the 3 trials per block). Recall that separate groups of participants completed the Magnified and Attenuated experimental conditions of the <italic>Variability</italic> task, and that, therefore, each group completed their own Control condition of the <italic>Variability</italic> task with no feedback perturbation. The Control versus experimental condition within-group comparisons in <xref ref-type="fig" rid="F8">Figure 8</xref> suggest that adaptation was not affected by the prior formant feedback variability manipulations. Statistical testing confirmed the absence of any significant differences between Control and Magnified or between Control and Attenuated for early adaptation extent (average formant frequency of the first 15 adaptation trials; <xref ref-type="fig" rid="F9">Figure 9A</xref>), learning rate during early adaptation (slope of a linear regression line over the formant frequencies of the first 15 adaptation trials; <xref ref-type="fig" rid="F8">Figure 8B</xref>), or final adaptation extent (average formant frequency of the last 15 perturbation trials; <xref ref-type="fig" rid="F8">Figure 8C</xref>). The <italic>p</italic> values for all statistical comparisons are included with the data visualizations in <xref ref-type="fig" rid="F9">Figure 9</xref>.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption><p>Group-level formant-shift adaptation data after completion of the <italic>Variability</italic> task&#x2019;s Control and Magnified conditions <bold>(A)</bold> or after the Control and Attenuated conditions <bold>(B)</bold>. Dots represent group mean formant frequencies per block (3 trials) and averaged across F1 and F2. Shaded regions indicate standard error of the mean. Solid lines are loess smoothed fits (span = 0.3).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-16-890065-g008.tif"/>
</fig>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption><p>Boxplots with symbols depicting each participant&#x2019;s early adaptation extent <bold>(A,B)</bold>, early adaptation rate <bold>(C,D)</bold>, and final adaptation extent <bold>(E,F)</bold> for formant-shift adaptation completed after the <italic>Variability</italic> task&#x2019;s Control and Magnified or Control and Attenuated conditions. Full statistics for these data: <bold>(A)</bold> <italic>t</italic>(13) = 0.366, <italic>p</italic> = 0.720, <italic>d</italic> = 0.098; <bold>(B)</bold> <italic>t</italic>(12) = &#x2013;0.822, <italic>p</italic> = 0.427, <italic>d</italic> = 0.228; <bold>(C)</bold> <italic>t</italic>(13) = 0.150, <italic>p</italic> = 0.883, <italic>d</italic> = 0.040; <bold>(D)</bold> <italic>t</italic>(13) = &#x2013;0.128, <italic>p</italic> = 0.900, <italic>d</italic> = 0.035; <bold>(E)</bold> <italic>t</italic>(13) = 0.301, <italic>p</italic> = 0.768, <italic>d</italic> = 0.081; <bold>(F)</bold> <italic>t</italic>(12) = &#x2013;0.450, <italic>p</italic> = 0.661, <italic>d</italic> = 0.125.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnhum-16-890065-g009.tif"/>
</fig>
</sec>
</sec>
<sec id="S4" sec-type="discussion">
<title>Discussion</title>
<p>Previous observational studies have led to the suggestion that inter-trial motor variability may be related to both enhanced online feedback-based compensation (a study on fundamental frequency in speech, <xref ref-type="bibr" rid="B42">Scheerer and Jones, 2012</xref>) and enhanced adaptive learning (a study on upper limb reach movements, <xref ref-type="bibr" rid="B56">Wu et al., 2014</xref>). However, neither of these results have been consistently supported by other empirical data (<xref ref-type="bibr" rid="B42">Scheerer and Jones, 2012</xref>; <xref ref-type="bibr" rid="B20">He et al., 2016</xref>; <xref ref-type="bibr" rid="B44">Singh et al., 2016</xref>), alternative explanations for the findings have been offered (<xref ref-type="bibr" rid="B20">He et al., 2016</xref>; <xref ref-type="bibr" rid="B44">Singh et al., 2016</xref>; <xref ref-type="bibr" rid="B30">Murillo et al., 2017</xref>; <xref ref-type="bibr" rid="B46">Sternad, 2018</xref>; <xref ref-type="bibr" rid="B52">van der Vliet et al., 2018</xref>), and further investigation is clearly warranted (<xref ref-type="bibr" rid="B16">Dhawale et al., 2017</xref>). Moreover, results from an experimental study that directly manipulated feedback variability for reaching movements by magnifying or attenuating the size of target errors suggested that the temporal structure of adjustments across trials, indexed by the sample lag 1 autocorrelation [ACF(1)] for movement endpoints, changed with manipulated feedback (<xref ref-type="bibr" rid="B49">van Beers, 2009</xref>). In the same study, the adjustments across trials were consistent with predictions made by state-space models often used to characterize learning mechanisms in sensorimotor adaptation experiments (<xref ref-type="bibr" rid="B49">van Beers, 2009</xref>). Thus, inter-trial motor variability itself may represent a form of trial-by-trial learning. On the other hand, the authors of a reaching movement experiment combining error feedback magnification or attenuation with a constant perturbation that elicits visuomotor adaptation concluded that variability manipulation did not alter the underlying adaptive learning mechanisms (<xref ref-type="bibr" rid="B51">van der Kooij et al., 2015</xref>), despite observed changes in adaptation behavior (<xref ref-type="bibr" rid="B36">Patton et al., 2013</xref>; <xref ref-type="bibr" rid="B51">van der Kooij et al., 2015</xref>).</p>
<p>We sought to clarify, for sensorimotor control of speech articulation, whether <italic>experimental manipulations</italic> of inter-trial feedback variability (here variability of formant frequencies in the real-time auditory feedback) (a) lead to speaker adjustments in inter-trial production variability, suggestive of an active regulation mechanism; (b) lead to changes in the temporal structure of adjustments across trials [ACF(1)], suggestive of trial-by-trial learning; and (c) affect learning in a subsequent auditory-motor adaptation paradigm with a constant formant-shift perturbation. To manipulate inter-trial formant variability in the feedback, we implemented a novel real-time formant manipulation algorithm that can either magnify or attenuate the difference between the formants in a current production and target formants operationally defined as the median formant values from a <italic>Pre-test</italic>.</p>
<sec id="S4.SS1">
<title>Active Regulation of Variability</title>
<p>After the <italic>Pre-test</italic> with unaltered auditory feedback, participants completed two conditions of a <italic>Variability</italic> task (each followed by an <italic>Adaptation</italic> task): one was a Control condition with unaltered formant feedback, and the other condition had either Magnified or Attenuated formant variability in the auditory feedback, depending on the participant&#x2019;s group assignment. Signal processing algorithms generating the feedback signal in these experimental conditions increased or decreased the distance between the formants produced in a given trial and the participants&#x2019; median formants for the same word in the <italic>Pre-test</italic>. We therefore quantified participants&#x2019; productions with a DI that expressed produced formant frequencies also in terms of their distance to the pre-test median.</p>
<p>Compared with each group&#x2019;s own Control condition, the condition with Magnified feedback variability did not result in an adjustment in distance, but the condition with Attenuated feedback variability led to a gradual <italic>increase</italic> in distance between produced trials and the pre-test median (thus opposing the feedback manipulation). This increasing distance between produced formants and pre-test median formants was detected in both the initial portion of the vowel (5&#x2013;30% into the total vowel duration; results in <xref ref-type="fig" rid="F4">Figure 4</xref>) and the middle portion of the vowel (40&#x2013;60% into the total vowel duration; results in <xref ref-type="fig" rid="F5">Figure 5</xref>) portions of the vowel, and, thus, reflects gradual changes in movement planning rather than online within-vowel corrections. In fact, neither of the experimental conditions affected the extent of within-vowel corrections as compared with the same participants&#x2019; Control condition. As it is theoretically possible for DI to increase even in the absence of an increase in variability (e.g., if a participant moved their formants further from the pre-test median but always to the same location in acoustic F1F2 space), we followed up by determining the size of the area in acoustic space covered by each participants&#x2019; productions. This analysis confirmed that during the early stages of exposure to Attenuated variability feedback, most&#x2014;but not all&#x2014;participants did actually increase the overall spread of their productions in the two-dimensional acoustic space (i.e., increased formant production variability; results in <xref ref-type="fig" rid="F6">Figure 6</xref>).</p>
<p>It is not straightforward to compare this finding of active variability regulation with those from prior limb motor control studies that magnified and/or attenuated the dispersion of feedback across trials as a by-product of manipulating the magnitude of target error in each trial. The study by <xref ref-type="bibr" rid="B55">Wong et al. (2009)</xref> only <italic>increased</italic> the size of perceived target errors (and thus feedback dispersion), and, consequently, one cannot necessarily attribute the resulting decrease in motor variability to the magnified feedback variability as opposed to a control strategy that seeks to avoid large errors on each trial individually. The study by <xref ref-type="bibr" rid="B49">van Beers (2009)</xref> did implement both magnified and attenuated target errors, but focused on the temporal structure of movement endpoint adjustments across trials (see below Section &#x201C;Temporal Structure&#x201D;). Nevertheless, for reaching movements with unperturbed visual feedback, <xref ref-type="bibr" rid="B49">van Beers (2009)</xref> reported that trial-to-trial adjustments are made in such a way that movement variability is minimized.</p>
<p>Our data from speech articulation are not consistent with the idea that the central nervous system generally aims to minimize variability. In fact, these data suggest a strikingly different situation: when the feedback perturbation magnified inter-trial formant variability, this extended variability was tolerated and not opposed, but when the perturbation attenuated inter-trial formant variability, articulation was gradually adjusted such that the acoustic output counteracted the perturbation. Thus, overall, the present data are consistent with the interpretation that a sufficiently large level of feedback variability is desirable, and that this level of variability is actively regulated through adjustments in motor planning.</p>
<p>In light of this overall support for the hypothesis that variability is actively regulated, it is reasonable to wonder why the increase in production variability in the Attenuated condition was not statistically significant in some of the later stages of the task. As shown in <xref ref-type="fig" rid="F4">Figure 4</xref>, the increase in DI<sub>initial</sub> relative to stage 1 was significant in stages 2 and 3 but not in stages 4 and 5 (in both cases <italic>p</italic> = 0.076 with medium effect sizes). Closer inspection reveals that, at Stage 4, the mean DI<sub>initial</sub> value had further increased, but the standard error of the mean was also larger at this stage. At Stage 5, the mean DI<sub>initial</sub> value did decrease, but it never returned to its original value from stage 1. As shown in <xref ref-type="fig" rid="F5">Figure 5</xref>, the increase in DI<sub>mid</sub> relative to stage 1 was still statistically significant in the last stage of the task, only not in the preceding Stage 4 (<italic>p</italic> = 0.078, medium effect size). Thus, there was a trend for the increased production variability to be not sustained at its maximum level in the later stages of the task, but any attempts at interpreting the specific results for Stage 4 would be purely speculative.</p>
<p>It should also be acknowledged that an alternative explanation might be offered for the absence of formant variability regulation in the Magnified condition of our <italic>Variability</italic> task. Specifically, one could argue that the highly practiced speech movements may have been performed with minimized variability from the very beginning of the task, and that, therefore, a floor effect prevented further reduction of this variability in the Magnified condition. This would be a reasonable argument as the lower bound of variability seems to be physiologically constrained by the stochastic nature of events in the peripheral motor system such as synaptic transmission (<xref ref-type="bibr" rid="B6">Calvin and Stevens, 1968</xref>) and muscle contraction (<xref ref-type="bibr" rid="B10">Clamann, 1969</xref>; <xref ref-type="bibr" rid="B19">Hamilton et al., 2004</xref>), which together are referred to as execution or performance noise in theoretical models of motor control (<xref ref-type="bibr" rid="B50">Van Beers et al., 2004</xref>; <xref ref-type="bibr" rid="B8">Cheng and Sabes, 2006</xref>; <xref ref-type="bibr" rid="B49">van Beers, 2009</xref>; <xref ref-type="bibr" rid="B16">Dhawale et al., 2017</xref>; <xref ref-type="bibr" rid="B52">van der Vliet et al., 2018</xref>). Only a separate component of motor variability, namely, planning or state noise (<xref ref-type="bibr" rid="B8">Cheng and Sabes, 2006</xref>; <xref ref-type="bibr" rid="B49">van Beers, 2009</xref>), may be subject to regulation by the central nervous system (<xref ref-type="bibr" rid="B56">Wu et al., 2014</xref>; <xref ref-type="bibr" rid="B16">Dhawale et al., 2017</xref>, <xref ref-type="bibr" rid="B15">2019</xref>). Total system noise always comprises both execution and planning noise, and, thus, cannot be regulated to a level lower than that of the execution noise itself. In fact, work on limb motor control has estimated the planning noise to be substantially smaller than the execution noise, the former accounting for only about 20&#x223C;30% of total motor variability (<xref ref-type="bibr" rid="B9">Cheng and Sabes, 2007</xref>; <xref ref-type="bibr" rid="B49">van Beers, 2009</xref>; <xref ref-type="bibr" rid="B52">van der Vliet et al., 2018</xref>).</p>
<p>However, the argument that the central nervous system does control speech movements in such a way that total system noise is minimized is not compatible with our results from the Attenuated condition. There would be no reason to implement adjustments in the direction of <italic>more</italic> variability in this condition if the controller seeks to minimize total system noise (given that the Attenuated feedback signal indicates a variability level that is minimized even below the presumed lower bound in typical speech). Consequently, our results from the two conditions taken together support the aforementioned interpretation that, at least for speech articulation, a certain non-minimal level of feedback variability is desirable and actively maintained, possibly in function of providing sensorimotor exploration (<xref ref-type="bibr" rid="B56">Wu et al., 2014</xref>; <xref ref-type="bibr" rid="B16">Dhawale et al., 2017</xref>, <xref ref-type="bibr" rid="B15">2019</xref>). Moreover, this conclusion implies that the speech motor control system not only calculates and keeps track of distribution features for key aspects of the auditory feedback signal (e.g., dispersion measures such as variance of the formant frequencies), but also compares these features with the expected distributions and then updates future movement planning accordingly (<xref ref-type="bibr" rid="B34">Parrell and Houde, 2019</xref>). If our findings are replicated in future studies, computational and conceptual models of speech motor control will need to start incorporating such more complex feedback mechanisms, analogous to suggestions that have been made in the non-speech motor control literature (e.g., <xref ref-type="bibr" rid="B21">Herzfeld et al., 2014</xref>; <xref ref-type="bibr" rid="B15">Dhawale et al., 2019</xref>).</p>
</sec>
<sec id="S4.SS2">
<title>Temporal Structure</title>
<p>If articulatory adjustments in the Attenuated condition of the <italic>Variability</italic> task relied on error-based learning mechanisms similar to those driving auditory-motor adaptation with predictable formant perturbations (<xref ref-type="bibr" rid="B23">Houde and Jordan, 1998</xref>; <xref ref-type="bibr" rid="B14">Daliri and Dittman, 2019</xref>), then the temporal structure of adjustments across trials&#x2014;such as indexed by the lag 1 autocorrelation [ACF(1)] of the overall sequence of productions&#x2014;would be expected to vary depending on the feedback manipulation (<xref ref-type="bibr" rid="B49">van Beers, 2009</xref>). It should be noted at this time that the authors of one previous publication on variability in formant production suggested that their ACF(1) results indicated trial-to-trial adjustments even for speech produced without any auditory perturbation (<xref ref-type="bibr" rid="B45">Sitek et al., 2013</xref>). However, the lag 1 autocorrelation of &#x2013;0.47 in that study was calculated based on <italic>differences</italic> between pairs of successive trials, thus introducing the problem of overdifferencing that we have discussed above in the Introduction (recall that after differencing even a white noise time series has a lag 1 autocorrelation of &#x2013;0.5). With regard to the specific perturbation-related questions investigated in the present study, our results (illustrated in <xref ref-type="fig" rid="F7">Figure 7</xref>) showed no statistically significant difference in ACF(1) for the sequences of trials produced in the conditions with Attenuated or Magnified formant feedback variability versus the Control condition with unaltered auditory feedback.</p>
<p>The lack of significant difference in ACF(1) between the Control and experimental conditions (Attenuated and Magnified) is not consistent with work by <xref ref-type="bibr" rid="B49">van Beers (2009)</xref>. In the latter study, comparisons with a Control condition showed that ACF(1) decreased in a Magnified condition and increased in an Attenuated condition, in keeping with the prediction of a state-space model of adaptive learning based on sensory feedback (<xref ref-type="bibr" rid="B8">Cheng and Sabes, 2006</xref>). In fact, in our own study, most of the ACF(1) values for the sequences of productions fell within the 95% confidence interval of a white noise process, suggesting no feedback-based learning. One possible interpretation is of course that the speech control system simply does not modify productions based on auditory feedback from the immediately preceding trial. Although this control system clearly shows adaptation to predictable auditory perturbations (<xref ref-type="bibr" rid="B23">Houde and Jordan, 1998</xref>; <xref ref-type="bibr" rid="B53">Villacorta et al., 2007</xref>; <xref ref-type="bibr" rid="B43">Shiller et al., 2020</xref>), it is possible that such learning mechanisms are inactive in the absence of consistently maintained predictable perturbations (cf. <xref ref-type="bibr" rid="B18">Gonzalez Castro et al., 2014</xref>; <xref ref-type="bibr" rid="B21">Herzfeld et al., 2014</xref>). However, the statistically significant increase in formant production variability in the Attenuated condition does indicate a previously undocumented form of adaptive learning process during this <italic>Variability</italic> task.</p>
<p>We therefore speculate that the employed ACF(1) analysis may fail to capture the specific form of feedback-based learning in the <italic>Variability</italic> task. Several observations support this hypothesis. First, the state-space model of motor control predicts that when the parameter of error sensitivity (also known as adaptation rate) is very low, the ACF(1) of the trial sequence for each of the feedback manipulations implemented in the current experiment would be small, and the trial sequence would resemble a white noise process (<xref ref-type="bibr" rid="B49">van Beers, 2009</xref>; <xref ref-type="bibr" rid="B52">van der Vliet et al., 2018</xref>). It has been estimated recently that, in comparison with limb motor control studies which generally reported error sensitivity in the range of 30&#x2013;50% (<xref ref-type="bibr" rid="B1">Baddeley et al., 2003</xref>; <xref ref-type="bibr" rid="B9">Cheng and Sabes, 2007</xref>; <xref ref-type="bibr" rid="B49">van Beers, 2009</xref>; <xref ref-type="bibr" rid="B51">van der Kooij et al., 2015</xref>), the error sensitivity for speech auditory-motor adaptation is, on average, as small as 4.8% (<xref ref-type="bibr" rid="B14">Daliri and Dittman, 2019</xref>). Second, it is known from previous studies that adaptive learning in speech production can differ between different vowels and words, and a given participant may even adapt for one vowel but follow the perturbation for another vowel (<xref ref-type="bibr" rid="B23">Houde and Jordan, 1998</xref>; <xref ref-type="bibr" rid="B29">Max and Maffett, 2015</xref>). In the current study&#x2019;s <italic>Variability</italic> task, three different target words (&#x201C;talk,&#x201D; &#x201C;tech,&#x201D; &#x201C;tuck&#x201D;) were produced in pseudo-random order. Feedback-based learning under such circumstances may be very complex (e.g., How much does feedback from a trial of &#x201C;tech&#x201D; affect the production of &#x201C;talk&#x201D;? What is the influence of some trials being preceded by the same word and other trials by a different word?), especially if the history of feedback prior to the last trial is also taken into account (<xref ref-type="bibr" rid="B21">Herzfeld et al., 2014</xref>). Such complexity is not captured by the simple ACF(1) index. Third, the statistically significant change in formant production variability during the Attenuated condition of the <italic>Variability</italic> task indicates that the production sequence may be non-stationary, which renders ACF(1) difficult to interpret. Unfortunately, despite these various disadvantages of ACF(1), it is unclear which alternative measurements may be used to reveal the temporal structure of feedback-based adaptive learning in conditions with altered formant feedback variability.</p>
</sec>
<sec id="S4.SS3">
<title>Effect of Variability on Adaptation</title>
<p>Immediately after having been exposed to Attenuated or Magnified formant feedback variability, participants completed a conventional speech auditory-motor adaptation task with a predictable upward shift of all formants. This <italic>Adaptation</italic> task allowed us to assess the potential effect of prior formant feedback variability on formant production learning. If sensorimotor learning is affected by the extent of perceived inter-trial variability (<xref ref-type="bibr" rid="B21">Herzfeld et al., 2014</xref>; <xref ref-type="bibr" rid="B56">Wu et al., 2014</xref>), then participants&#x2019; formant adaptation profiles can be expected to differ after experiencing Attenuated versus Magnified formant feedback variability. On the other hand, if inter-trial variability has no effect on the mechanisms underlying adaptive learning (<xref ref-type="bibr" rid="B51">van der Kooij et al., 2015</xref>), then participants&#x2019; formant adaptation profiles can be expected to remain unchanged between conditions.</p>
<p>Results shown in <xref ref-type="fig" rid="F8">Figures 8</xref>, <xref ref-type="fig" rid="F9">9</xref> indicate that three different measures of formant adaptation&#x2014;early adaptation extent, early adaptation rate, and final adaptation extent&#x2014;were all statistically indistinguishable between the Control condition and the two experimental conditions (Magnified or Attenuated feedback variability). In other words, the prior manipulation of formant feedback variability, or the participants&#x2019; motor adjustments to this manipulation, had no effect at all on the subsequent formant shift adaptation task. This result aligns with the conclusion of <xref ref-type="bibr" rid="B51">van der Kooij et al. (2015)</xref> who conducted a series of visuomotor rotation reach experiments with magnified or attenuated visual feedback errors. Although those authors observed behavioral differences across the feedback manipulation conditions, state-space model estimates of the underlying learning <italic>mechanism</italic> remained unchanged. In the current study, even the behavioral measures showed no differences at all in each group&#x2019;s comparison of formant adaptation after the control versus experimental condition of the <italic>Variability</italic> task. Hence, our results for speech articulation suggest no direct relationship between formant variability perceived in a preceding task and the adaptive learning of formant output adjustments when subsequently exposed to a persistent formant perturbation.</p>
<p>The absence of an effect of formant feedback variability on formant production adaptation may relate to the aforementioned low error-sensitivity parameter in speech auditory-motor adaptation (<xref ref-type="bibr" rid="B14">Daliri and Dittman, 2019</xref>). Of course, it is also possible that this outcome is entirely specific to certain methodological aspects of our study. For example, we only implemented a relatively <italic>short-term</italic> feedback variability manipulation (75 trials), and examined formant-shift adaptation in a <italic>subsequent</italic> task. Future studies should also address the effect of longer-term variability manipulations and variability manipulations implemented during the auditory-motor adaptation task itself. Moreover, it might prove fruitful to develop methodological approaches that are able to dissociate the effects of manipulations that alter sensory variability (as implemented here) versus direct manipulations of motor variability (which alter both motor and sensory variability).</p>
</sec>
</sec>
<sec id="S5" sec-type="conclusion">
<title>Conclusion</title>
<p>In sum, by experimentally manipulating inter-trial formant variability in the auditory feedback signal for speech, the present study yielded three novel findings. First, formant production variability in speech production appears to be actively regulated to a desirable level rather than merely minimized. Second, under the conditions investigated here, the temporal structure of inter-trial formant changes was not affected by experimental manipulations of formant feedback variability. Third, for these specific test conditions, subsequent auditory-motor adaptation in a standard formant shift perturbation task was also not affected by the formant feedback manipulations. We hope that future empirical studies will be able to investigate the generalizability of these findings, and that future theoretical work will provide conceptual and computational accounts of the active regulation of inter-trial variability in the sensorimotor control of speech production.</p>
</sec>
<sec id="S6" sec-type="data-availability">
<title>Data Availability Statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec id="S7">
<title>Ethics Statement</title>
<p>The studies involving human participants were reviewed and approved by University of Washington IRB. The participants provided their written informed consent to participate in this study.</p>
</sec>
<sec id="S8">
<title>Author Contributions</title>
<p>HW collected and analyzed the data. Both authors designed the experiments and data analysis procedures, interpreted the data, wrote the manuscript, contributed to the article, and approved the submitted version.</p>
</sec>
<sec id="audiscl1">
<title>Author Disclaimer</title>
<p>The content is solely the responsibility of the authors and does not necessarily represent the official views of theNational Institute on Deafness and Other Communication Disorders or the National Institutes of Health.</p>
</sec>
<sec id="conf1" sec-type="COI-statement">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="pudiscl1" sec-type="disclaimer">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
</body>
<back>
<sec id="S9" sec-type="funding-information">
<title>Funding</title>
<p>This research was supported by grants R01DC014510 and R01DC017444 from the National Institute on Deafness and Other Communication Disorders.</p>
</sec>
<ack><p>We thank Kwang S. Kim, Ph.D., for contributions during the process of designing and implementing the auditory feedback perturbations.</p>
</ack>
<ref-list>
<title>References</title>
<ref id="B1"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Baddeley</surname> <given-names>R. J.</given-names></name> <name><surname>Ingram</surname> <given-names>H. A.</given-names></name> <name><surname>Miall</surname> <given-names>R. C.</given-names></name></person-group> (<year>2003</year>). <article-title>System identification applied to a visuomotor task: near-optimal human performance in a noisy changing task.</article-title> <source><italic>J. Neurosci.</italic></source> <volume>23</volume> <fpage>3066</fpage>&#x2013;<lpage>3075</lpage>. <pub-id pub-id-type="doi">10.1523/jneurosci.23-07-03066.2003</pub-id> <pub-id pub-id-type="pmid">12684493</pub-id></citation></ref>
<ref id="B2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bieniek</surname> <given-names>M. M.</given-names></name> <name><surname>Bennett</surname> <given-names>P. J.</given-names></name> <name><surname>Sekuler</surname> <given-names>A. B.</given-names></name> <name><surname>Rousselet</surname> <given-names>G. A.</given-names></name></person-group> (<year>2016</year>). <article-title>A robust and representative lower bound on object processing speed in humans.</article-title> <source><italic>Eur. J. Neurosci.</italic></source> <volume>44</volume> <fpage>1804</fpage>&#x2013;<lpage>1814</lpage>. <pub-id pub-id-type="doi">10.1111/ejn.13100</pub-id> <pub-id pub-id-type="pmid">26469359</pub-id></citation></ref>
<ref id="B3"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Boersma</surname> <given-names>P.</given-names></name></person-group> (<year>2001</year>). <article-title>Praat, a system for doing phonetics by computer.</article-title> <source><italic>Glot Int.</italic></source> <volume>5</volume> <fpage>341</fpage>&#x2013;<lpage>345</lpage>.</citation></ref>
<ref id="B4"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brockwell</surname> <given-names>P. J.</given-names></name> <name><surname>Davis</surname> <given-names>R. A.</given-names></name></person-group> (<year>2016</year>). <source><italic>Introduction to Time Series and Forecasting.</italic></source> <publisher-loc>Berlin</publisher-loc>: <publisher-name>Springer</publisher-name>.</citation></ref>
<ref id="B5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cai</surname> <given-names>S.</given-names></name> <name><surname>Boucek</surname> <given-names>M.</given-names></name> <name><surname>Ghosh</surname> <given-names>S.</given-names></name> <name><surname>Guenther</surname> <given-names>F.</given-names></name> <name><surname>Perkell</surname> <given-names>J.</given-names></name></person-group> (<year>2008</year>). &#x201C;<article-title>A system for online dynamic perturbation of formant trajectories and results from perturbations of the mandarin triphthong /iau/</article-title>,&#x201D;in <source><italic>Proceedings of the 8th ISSP</italic></source>, <publisher-name>University of Strasbourg</publisher-name>, <publisher-loc>Strasbourg</publisher-loc>.</citation></ref>
<ref id="B6"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Calvin</surname> <given-names>W. H.</given-names></name> <name><surname>Stevens</surname> <given-names>C. F.</given-names></name></person-group> (<year>1968</year>). <article-title>Synaptic noise and other sources of randomness in motoneuron interspike intervals.</article-title> <source><italic>J. Neurophysiol.</italic></source> <volume>31</volume> <fpage>574</fpage>&#x2013;<lpage>587</lpage>. <pub-id pub-id-type="doi">10.1152/jn.1968.31.4.574</pub-id> <pub-id pub-id-type="pmid">5709873</pub-id></citation></ref>
<ref id="B7"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chao</surname> <given-names>S.-C.</given-names></name> <name><surname>Ochoa</surname> <given-names>D.</given-names></name> <name><surname>Daliri</surname> <given-names>A.</given-names></name></person-group> (<year>2019</year>). <article-title>Production variability and categorical perception of vowels are strongly linked.</article-title> <source><italic>Front. Hum. Neurosci.</italic></source> <volume>13</volume>:<issue>96</issue>. <pub-id pub-id-type="doi">10.3389/fnhum.2019.00096</pub-id> <pub-id pub-id-type="pmid">30967768</pub-id></citation></ref>
<ref id="B8"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cheng</surname> <given-names>S.</given-names></name> <name><surname>Sabes</surname> <given-names>P. N.</given-names></name></person-group> (<year>2006</year>). <article-title>Modeling sensorimotor learning with linear dynamical systems.</article-title> <source><italic>Neural Comput.</italic></source> <volume>18</volume> <fpage>760</fpage>&#x2013;<lpage>793</lpage>. <pub-id pub-id-type="doi">10.1162/089976606775774651</pub-id> <pub-id pub-id-type="pmid">16494690</pub-id></citation></ref>
<ref id="B9"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cheng</surname> <given-names>S.</given-names></name> <name><surname>Sabes</surname> <given-names>P. N.</given-names></name></person-group> (<year>2007</year>). <article-title>Calibration of visually guided reaching is driven by error-corrective learning and internal dynamics.</article-title> <source><italic>J. Neurophysiol.</italic></source> <volume>97</volume> <fpage>3057</fpage>&#x2013;<lpage>3069</lpage>. <pub-id pub-id-type="doi">10.1152/jn.00897.2006</pub-id> <pub-id pub-id-type="pmid">17202230</pub-id></citation></ref>
<ref id="B10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Clamann</surname> <given-names>H. P.</given-names></name></person-group> (<year>1969</year>). <article-title>Statistical analysis of motor unit firing patterns in a human skeletal muscle.</article-title> <source><italic>Biophys. J.</italic></source> <volume>9</volume> <fpage>1233</fpage>&#x2013;<lpage>1251</lpage>. <pub-id pub-id-type="doi">10.1016/S0006-3495(69)86448-9</pub-id></citation></ref>
<ref id="B11"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cohen</surname> <given-names>J.</given-names></name></person-group> (<year>1988</year>). <source><italic>Statistical Power Analysis for the Behavioral Sciences.</italic></source> <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Routledge Academic</publisher-name>.</citation></ref>
<ref id="B12"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cornelisse</surname> <given-names>L. E.</given-names></name> <name><surname>Gagne</surname> <given-names>J. P.</given-names></name> <name><surname>Seewald</surname> <given-names>R. C.</given-names></name></person-group> (<year>1991</year>). <article-title>Ear level recordings of the long-term average spectrum of speech.</article-title> <source><italic>Ear Hear.</italic></source> <volume>12</volume> <fpage>47</fpage>&#x2013;<lpage>54</lpage>. <pub-id pub-id-type="doi">10.1097/00003446-199102000-00006</pub-id> <pub-id pub-id-type="pmid">2026288</pub-id></citation></ref>
<ref id="B13"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cryer</surname> <given-names>J. D.</given-names></name> <name><surname>Chan</surname> <given-names>K.-S.</given-names></name></person-group> (<year>2008</year>). <source><italic>Time Series Analysis: With Applications to R.</italic></source> <publisher-loc>Berlin</publisher-loc>: <publisher-name>Springer</publisher-name>. <pub-id pub-id-type="doi">10.1007/978-0-387-75959-3</pub-id></citation></ref>
<ref id="B14"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Daliri</surname> <given-names>A.</given-names></name> <name><surname>Dittman</surname> <given-names>J.</given-names></name></person-group> (<year>2019</year>). <article-title>Successful auditory motor adaptation requires task-relevant auditory errors.</article-title> <source><italic>J. Neurophysiol.</italic></source> <volume>122</volume> <fpage>552</fpage>&#x2013;<lpage>562</lpage>. <pub-id pub-id-type="doi">10.1152/jn.00662.2018</pub-id> <pub-id pub-id-type="pmid">31215301</pub-id></citation></ref>
<ref id="B15"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dhawale</surname> <given-names>A. K.</given-names></name> <name><surname>Miyamoto</surname> <given-names>Y. R.</given-names></name> <name><surname>Smith</surname> <given-names>M. A.</given-names></name> <name><surname>&#x00D6;lveczky</surname> <given-names>B. P.</given-names></name></person-group> (<year>2019</year>). <article-title>Adaptive regulation of motor variability.</article-title> <source><italic>Curr. Biol.</italic></source> <volume>29</volume> <fpage>3551.e7</fpage>&#x2013;<lpage>3562.e7</lpage>. <pub-id pub-id-type="doi">10.1016/j.cub.2019.08.052</pub-id> <pub-id pub-id-type="pmid">31630947</pub-id></citation></ref>
<ref id="B16"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dhawale</surname> <given-names>A. K.</given-names></name> <name><surname>Smith</surname> <given-names>M. A.</given-names></name> <name><surname>&#x00D6;lveczky</surname> <given-names>B. P.</given-names></name></person-group> (<year>2017</year>). <article-title>The role of variability in motor learning.</article-title> <source><italic>Ann. Rev. Neurosci.</italic></source> <volume>40</volume> <fpage>479</fpage>&#x2013;<lpage>498</lpage>. <pub-id pub-id-type="doi">10.1146/annurev-neuro-072116-031548</pub-id> <pub-id pub-id-type="pmid">28489490</pub-id></citation></ref>
<ref id="B17"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Franken</surname> <given-names>M. K.</given-names></name> <name><surname>Acheson</surname> <given-names>D. J.</given-names></name> <name><surname>McQueen</surname> <given-names>J. M.</given-names></name> <name><surname>Eisner</surname> <given-names>F.</given-names></name> <name><surname>Hagoort</surname> <given-names>P.</given-names></name></person-group> (<year>2017</year>). <article-title>Individual variability as a window on production-perception interactions in speech motor control.</article-title> <source><italic>J. Acoust. Soc. Am.</italic></source> <volume>142</volume> <fpage>2007</fpage>&#x2013;<lpage>2018</lpage>. <pub-id pub-id-type="doi">10.1121/1.5006899</pub-id></citation></ref>
<ref id="B18"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gonzalez Castro</surname> <given-names>L. N.</given-names></name> <name><surname>Hadjiosif</surname> <given-names>A. M.</given-names></name> <name><surname>Hemphill</surname> <given-names>M. A.</given-names></name> <name><surname>Smith</surname> <given-names>M. A.</given-names></name></person-group> (<year>2014</year>). <article-title>Environmental consistency determines the rate of motor adaptation.</article-title> <source><italic>Curr. Biol.</italic></source> <volume>24</volume> <fpage>1050</fpage>&#x2013;<lpage>1061</lpage>. <pub-id pub-id-type="doi">10.1016/j.cub.2014.03.049</pub-id> <pub-id pub-id-type="pmid">24794296</pub-id></citation></ref>
<ref id="B19"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hamilton</surname> <given-names>A. F. D. C.</given-names></name> <name><surname>Jones</surname> <given-names>K. E.</given-names></name> <name><surname>Wolpert</surname> <given-names>D. M.</given-names></name></person-group> (<year>2004</year>). <article-title>The scaling of motor noise with muscle strength and motor unit number in humans.</article-title> <source><italic>Exp. Brain Res.</italic></source> <volume>157</volume> <fpage>417</fpage>&#x2013;<lpage>430</lpage>. <pub-id pub-id-type="doi">10.1007/s00221-004-1856-7</pub-id> <pub-id pub-id-type="pmid">15014922</pub-id></citation></ref>
<ref id="B20"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>He</surname> <given-names>K.</given-names></name> <name><surname>Liang</surname> <given-names>Y.</given-names></name> <name><surname>Abdollahi</surname> <given-names>F.</given-names></name> <name><surname>Fisher Bittmann</surname> <given-names>M.</given-names></name> <name><surname>Kording</surname> <given-names>K.</given-names></name> <name><surname>Wei</surname> <given-names>K.</given-names></name></person-group> (<year>2016</year>). <article-title>The statistical determinants of the speed of motor learning.</article-title> <source><italic>PLoS Comput. Biol.</italic></source> <volume>12</volume>:<issue>e1005023</issue>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1005023</pub-id> <pub-id pub-id-type="pmid">27606808</pub-id></citation></ref>
<ref id="B21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Herzfeld</surname> <given-names>D. J.</given-names></name> <name><surname>Vaswani</surname> <given-names>P. A.</given-names></name> <name><surname>Marko</surname> <given-names>M. K.</given-names></name> <name><surname>Shadmehr</surname> <given-names>R.</given-names></name></person-group> (<year>2014</year>). <article-title>A memory of errors in sensorimotor learning.</article-title> <source><italic>Science</italic></source> <volume>345</volume> <fpage>1349</fpage>&#x2013;<lpage>1353</lpage>. <pub-id pub-id-type="doi">10.1126/science.1253138</pub-id> <pub-id pub-id-type="pmid">25123484</pub-id></citation></ref>
<ref id="B22"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Holm</surname> <given-names>S.</given-names></name></person-group> (<year>1979</year>). <article-title>A simple sequentially rejective multiple test procedure.</article-title> <source><italic>Scand. J. Stat.</italic></source> <volume>6</volume> <fpage>65</fpage>&#x2013;<lpage>70</lpage>.</citation></ref>
<ref id="B23"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Houde</surname> <given-names>J. F.</given-names></name> <name><surname>Jordan</surname> <given-names>M. I.</given-names></name></person-group> (<year>1998</year>). <article-title>Sensorimotor adaptation in speech production.</article-title> <source><italic>Science</italic></source> <volume>279</volume> <fpage>1213</fpage>&#x2013;<lpage>1216</lpage>. <pub-id pub-id-type="doi">10.1126/science.279.5354.1213</pub-id> <pub-id pub-id-type="pmid">9469813</pub-id></citation></ref>
<ref id="B24"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kim</surname> <given-names>K. S.</given-names></name> <name><surname>Daliri</surname> <given-names>A.</given-names></name> <name><surname>Flanagan</surname> <given-names>J. R.</given-names></name> <name><surname>Max</surname> <given-names>L.</given-names></name></person-group> (<year>2020a</year>). <article-title>Dissociated development of speech and limb sensorimotor learning in stuttering: speech auditory-motor learning is impaired in both children and adults who stutter.</article-title> <source><italic>Neuroscience</italic></source> <volume>451</volume> <fpage>1</fpage>&#x2013;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroscience.2020.10.014</pub-id> <pub-id pub-id-type="pmid">33091464</pub-id></citation></ref>
<ref id="B25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kim</surname> <given-names>K. S.</given-names></name> <name><surname>Wang</surname> <given-names>H.</given-names></name> <name><surname>Max</surname> <given-names>L.</given-names></name></person-group> (<year>2020b</year>). <article-title>It&#x2019;s about time: minimizing hardware and software latencies in speech research with real-time auditory feedback.</article-title> <source><italic>J. Speech Lang. Hear. Res.</italic></source> <volume>63</volume> <fpage>2522</fpage>&#x2013;<lpage>2534</lpage>. <pub-id pub-id-type="doi">10.1044/2020_JSLHR-19-00419</pub-id></citation></ref>
<ref id="B26"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lindblom</surname> <given-names>B.</given-names></name></person-group> (<year>1990</year>). <article-title>Explaining phonetic variation: a sketch of the H&#x0026;H theory.</article-title> <source><italic>Speech Prod. Speech Model.</italic></source> <volume>55</volume> <fpage>403</fpage>&#x2013;<lpage>439</lpage>. <pub-id pub-id-type="doi">10.1007/978-94-009-2037-8_16</pub-id></citation></ref>
<ref id="B27"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>MacDonald</surname> <given-names>E. N.</given-names></name> <name><surname>Purcell</surname> <given-names>D. W.</given-names></name> <name><surname>Munhall</surname> <given-names>K. G.</given-names></name></person-group> (<year>2011</year>). <article-title>Probing the independence of formant control using altered auditory feedback.</article-title> <source><italic>J. Acoust. Soc. Am.</italic></source> <volume>129</volume>:<issue>955</issue>. <pub-id pub-id-type="doi">10.1121/1.3531932</pub-id></citation></ref>
<ref id="B28"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>MacNeilage</surname> <given-names>P. F.</given-names></name></person-group> (<year>1970</year>). <article-title>Motor control of serial ordering of speech.</article-title> <source><italic>Psychol. Rev.</italic></source> <volume>77</volume> <fpage>182</fpage>&#x2013;<lpage>196</lpage>. <pub-id pub-id-type="doi">10.1037/h0029070</pub-id> <pub-id pub-id-type="pmid">5454132</pub-id></citation></ref>
<ref id="B29"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Max</surname> <given-names>L.</given-names></name> <name><surname>Maffett</surname> <given-names>D. G.</given-names></name></person-group> (<year>2015</year>). <article-title>Feedback delays eliminate auditory-motor learning in speech production.</article-title> <source><italic>Neurosci. Lett.</italic></source> <volume>591</volume> <fpage>25</fpage>&#x2013;<lpage>29</lpage>.</citation></ref>
<ref id="B30"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Murillo</surname> <given-names>D. B.</given-names></name> <name><surname>S&#x00E1;nchez</surname> <given-names>C. C.</given-names></name> <name><surname>Moreside</surname> <given-names>J.</given-names></name> <name><surname>Vera-Garc&#x00ED;a</surname> <given-names>F. J.</given-names></name> <name><surname>Moreno</surname> <given-names>F. J.</given-names></name></person-group> (<year>2017</year>). <article-title>Can the structure of motor variability predict learning rate?</article-title> <source><italic>J. Exp. Psychol.</italic></source> <volume>43</volume> <fpage>596</fpage>&#x2013;<lpage>607</lpage>. <pub-id pub-id-type="doi">10.1037/xhp0000303</pub-id> <pub-id pub-id-type="pmid">28095006</pub-id></citation></ref>
<ref id="B31"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nault</surname> <given-names>D. R.</given-names></name> <name><surname>Munhall</surname> <given-names>K. G.</given-names></name></person-group> (<year>2020</year>). <article-title>Individual variability in auditory feedback processing: responses to real-time formant perturbations and their relation to perceptual acuity.</article-title> <source><italic>J. Acoust. Soc. Am.</italic></source> <volume>148</volume>:<issue>3709</issue>. <pub-id pub-id-type="doi">10.1121/10.0002923</pub-id></citation></ref>
<ref id="B32"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Niziolek</surname> <given-names>C. A.</given-names></name> <name><surname>Kiran</surname> <given-names>S.</given-names></name></person-group> (<year>2018</year>). <article-title>Assessing speech correction abilities with acoustic analyses: evidence of preserved online correction in persons with aphasia.</article-title> <source><italic>Int. J. Speech Lang. Pathol.</italic></source> <volume>20</volume> <fpage>659</fpage>&#x2013;<lpage>668</lpage>. <pub-id pub-id-type="doi">10.1080/17549507.2018.1498920</pub-id> <pub-id pub-id-type="pmid">30348017</pub-id></citation></ref>
<ref id="B33"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Niziolek</surname> <given-names>C. A.</given-names></name> <name><surname>Nagarajan</surname> <given-names>S. S.</given-names></name> <name><surname>Houde</surname> <given-names>J. F.</given-names></name></person-group> (<year>2013</year>). <article-title>What does motor efference copy represent? evidence from speech production.</article-title> <source><italic>J. Neurosci.</italic></source> <volume>33</volume> <fpage>16110</fpage>&#x2013;<lpage>16116</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.2137-13.2013</pub-id> <pub-id pub-id-type="pmid">24107944</pub-id></citation></ref>
<ref id="B34"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Parrell</surname> <given-names>B.</given-names></name> <name><surname>Houde</surname> <given-names>J.</given-names></name></person-group> (<year>2019</year>). <article-title>Modeling the role of sensory feedback in speech motor control and learning.</article-title> <source><italic>J. Speech Lang. Hear. Res.</italic></source> <volume>62</volume> <fpage>2963</fpage>&#x2013;<lpage>2985</lpage>. <pub-id pub-id-type="doi">10.1044/2019_JSLHR-S-CSMC7-18-0127</pub-id></citation></ref>
<ref id="B35"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Patri</surname> <given-names>J. F.</given-names></name> <name><surname>Diard</surname> <given-names>J.</given-names></name> <name><surname>Perrier</surname> <given-names>P.</given-names></name></person-group> (<year>2015</year>). <article-title>Optimal speech motor control and token-to-token variability: a Bayesian modeling approach.</article-title> <source><italic>Biol. Cybernet.</italic></source> <volume>109</volume> <fpage>611</fpage>&#x2013;<lpage>626</lpage>. <pub-id pub-id-type="doi">10.1007/s00422-015-0664-4</pub-id> <pub-id pub-id-type="pmid">26497359</pub-id></citation></ref>
<ref id="B36"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Patton</surname> <given-names>J. L.</given-names></name> <name><surname>Wei</surname> <given-names>Y. J.</given-names></name> <name><surname>Bajaj</surname> <given-names>P.</given-names></name> <name><surname>Scheidt</surname> <given-names>R. A.</given-names></name></person-group> (<year>2013</year>). <article-title>Visuomotor learning enhanced by augmenting instantaneous trajectory error feedback during reaching.</article-title> <source><italic>PLoS One</italic></source> <volume>8</volume>:<issue>e46466</issue>. <pub-id pub-id-type="doi">10.1371/journal.pone.0046466</pub-id> <pub-id pub-id-type="pmid">23382796</pub-id></citation></ref>
<ref id="B37"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Perkell</surname> <given-names>J. S.</given-names></name> <name><surname>Klatt</surname> <given-names>D. H.</given-names></name></person-group> (<year>1986</year>). <source><italic>Invariance and Variability in Speech Processes.</italic></source> <publisher-loc>Mahwah, NJ</publisher-loc>: <publisher-name>Lawrence Erlbaum Associates</publisher-name>.</citation></ref>
<ref id="B38"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Perkell</surname> <given-names>J. S.</given-names></name> <name><surname>Lane</surname> <given-names>H.</given-names></name> <name><surname>Ghosh</surname> <given-names>S.</given-names></name> <name><surname>Matthies</surname> <given-names>M. L.</given-names></name> <name><surname>Tiede</surname> <given-names>M.</given-names></name> <name><surname>Guenther</surname> <given-names>F.</given-names></name><etal/></person-group> (<year>2008</year>). &#x201C;<article-title>Mechanisms of vowel production: auditory goals and speaker acuity</article-title>,&#x201D; in <source><italic>Proceeding of the Paper Presented at the 8th International Seminar on Speech Production</italic></source>, <publisher-loc>Groningen</publisher-loc>. <pub-id pub-id-type="pmid">17902866</pub-id></citation></ref>
<ref id="B39"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Purcell</surname> <given-names>D. W.</given-names></name> <name><surname>Munhall</surname> <given-names>K. G.</given-names></name></person-group> (<year>2006</year>). <article-title>Adaptive control of vowel formant frequency: evidence from real-time formant manipulation.</article-title> <source><italic>J. Acoust. Soc. Am.</italic></source> <volume>120</volume> <fpage>966</fpage>&#x2013;<lpage>977</lpage>. <pub-id pub-id-type="doi">10.1121/1.2217714</pub-id></citation></ref>
<ref id="B40"><citation citation-type="journal"><collab>R Core Team</collab> (<year>2019</year>). <source><italic>R: A Language and Environment for Statistical Computing</italic></source>. <comment>R Foundation for Statistical Computing</comment>.</citation></ref>
<ref id="B41"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rousselet</surname> <given-names>G. A.</given-names></name> <name><surname>Pernet</surname> <given-names>C. R.</given-names></name> <name><surname>Wilcox</surname> <given-names>R. R.</given-names></name></person-group> (<year>2017</year>). <article-title>Beyond differences in means: robust graphical methods to compare two groups in neuroscience.</article-title> <source><italic>Eur. J. Neurosci.</italic></source> <volume>46</volume> <fpage>1738</fpage>&#x2013;<lpage>1748</lpage>. <pub-id pub-id-type="doi">10.1111/ejn.13610</pub-id> <pub-id pub-id-type="pmid">28544058</pub-id></citation></ref>
<ref id="B42"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Scheerer</surname> <given-names>N. E.</given-names></name> <name><surname>Jones</surname> <given-names>J. A.</given-names></name></person-group> (<year>2012</year>). <article-title>The relationship between vocal accuracy and variability to the level of compensation to altered auditory feedback.</article-title> <source><italic>Neurosci. Lett.</italic></source> <volume>529</volume> <fpage>128</fpage>&#x2013;<lpage>132</lpage>. <pub-id pub-id-type="doi">10.1016/j.neulet.2012.09.012</pub-id> <pub-id pub-id-type="pmid">22995182</pub-id></citation></ref>
<ref id="B43"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shiller</surname> <given-names>D. M.</given-names></name> <name><surname>Mitsuya</surname> <given-names>T.</given-names></name> <name><surname>Max</surname> <given-names>L.</given-names></name></person-group> (<year>2020</year>). <article-title>Exposure to auditory feedback delay while speaking induces perceptual habituation but does not mitigate the disruptive effect of delay on speech auditory-motor learning.</article-title> <source><italic>Neuroscience</italic></source> <volume>446</volume> <fpage>213</fpage>&#x2013;<lpage>224</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroscience.2020.07.041</pub-id> <pub-id pub-id-type="pmid">32738430</pub-id></citation></ref>
<ref id="B44"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Singh</surname> <given-names>P.</given-names></name> <name><surname>Jana</surname> <given-names>S.</given-names></name> <name><surname>Ghosal</surname> <given-names>A.</given-names></name> <name><surname>Murthy</surname> <given-names>A.</given-names></name> <name><surname>Goldberg</surname> <given-names>M. E.</given-names></name></person-group> (<year>2016</year>). <article-title>Exploration of joint redundancy but not task space variability facilitates supervised motor learning.</article-title> <source><italic>Proc. Natl. Acad. Sci. U.S.A.</italic></source> <volume>113</volume> <fpage>14414</fpage>&#x2013;<lpage>14419</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.1613383113</pub-id> <pub-id pub-id-type="pmid">27911808</pub-id></citation></ref>
<ref id="B45"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sitek</surname> <given-names>K. R.</given-names></name> <name><surname>Mathalon</surname> <given-names>D. H.</given-names></name> <name><surname>Roach</surname> <given-names>B. J.</given-names></name> <name><surname>Houde</surname> <given-names>J. F.</given-names></name> <name><surname>Niziolek</surname> <given-names>C. A.</given-names></name> <name><surname>Ford</surname> <given-names>J. M.</given-names></name></person-group> (<year>2013</year>). <article-title>Auditory cortex processes variation in our own speech.</article-title> <source><italic>PLoS One</italic></source> <volume>8</volume>:<issue>e82925</issue>. <pub-id pub-id-type="doi">10.1371/journal.pone.0082925</pub-id> <pub-id pub-id-type="pmid">24349399</pub-id></citation></ref>
<ref id="B46"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sternad</surname> <given-names>D.</given-names></name></person-group> (<year>2018</year>). <article-title>It&#x2019;s not (only) the mean that matters: variability, noise and exploration in skill learning.</article-title> <source><italic>Curr. Opin. Behav. Sci.</italic></source> <volume>20</volume> <fpage>183</fpage>&#x2013;<lpage>195</lpage>. <pub-id pub-id-type="doi">10.1016/j.cobeha.2018.01.004</pub-id> <pub-id pub-id-type="pmid">30035207</pub-id></citation></ref>
<ref id="B47"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tang</surname> <given-names>D.</given-names></name> <name><surname>Parrell</surname> <given-names>B.</given-names></name> <name><surname>Niziolek</surname> <given-names>C. A.</given-names></name></person-group> (<year>2021</year>). <article-title>Variability is actively regulated in speech.</article-title> <source><italic>bioRxiv</italic></source> [<comment>Preprint</comment>]. <pub-id pub-id-type="doi">10.1101/2021.10.08.462639</pub-id></citation></ref>
<ref id="B48"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tourville</surname> <given-names>J. A.</given-names></name> <name><surname>Cai</surname> <given-names>S.</given-names></name> <name><surname>Guenther</surname> <given-names>F.</given-names></name></person-group> (<year>2013</year>). <article-title>Exploring auditory-motor interactions in normal and disordered speech.</article-title> <source><italic>Proc. Meet. Acoust.</italic></source> <volume>19</volume>:<issue>060180</issue>. <pub-id pub-id-type="doi">10.1121/1.4800684</pub-id></citation></ref>
<ref id="B49"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>van Beers</surname> <given-names>R. J.</given-names></name></person-group> (<year>2009</year>). <article-title>Motor learning is optimally tuned to the properties of motor noise.</article-title> <source><italic>Neuron</italic></source> <volume>63</volume> <fpage>406</fpage>&#x2013;<lpage>417</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2009.06.025</pub-id> <pub-id pub-id-type="pmid">19679079</pub-id></citation></ref>
<ref id="B50"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van Beers</surname> <given-names>R. J.</given-names></name> <name><surname>Haggard</surname> <given-names>P.</given-names></name> <name><surname>Wolpert</surname> <given-names>D. M.</given-names></name></person-group> (<year>2004</year>). <article-title>The role of execution noise in movement variability.</article-title> <source><italic>J. Neurophysiol.</italic></source> <volume>91</volume> <fpage>1050</fpage>&#x2013;<lpage>1063</lpage>. <pub-id pub-id-type="doi">10.1152/jn.00652.2003</pub-id> <pub-id pub-id-type="pmid">14561687</pub-id></citation></ref>
<ref id="B51"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>van der Kooij</surname> <given-names>K.</given-names></name> <name><surname>Brenner</surname> <given-names>E.</given-names></name> <name><surname>van Beers</surname> <given-names>R. J.</given-names></name> <name><surname>Smeets</surname> <given-names>J. B. J.</given-names></name></person-group> (<year>2015</year>). <article-title>Visuomotor adaptation: how forgetting keeps us conservative.</article-title> <source><italic>PLoS One</italic></source> <volume>10</volume>:<issue>e0117901</issue>. <pub-id pub-id-type="doi">10.1371/journal.pone.0117901</pub-id> <pub-id pub-id-type="pmid">25723763</pub-id></citation></ref>
<ref id="B52"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>van der Vliet</surname> <given-names>R.</given-names></name> <name><surname>Frens</surname> <given-names>M. A.</given-names></name> <name><surname>de Vreede</surname> <given-names>L.</given-names></name> <name><surname>Jonker</surname> <given-names>Z. D.</given-names></name> <name><surname>Ribbers</surname> <given-names>G. M.</given-names></name> <name><surname>Selles</surname> <given-names>R. W.</given-names></name><etal/></person-group> (<year>2018</year>). <article-title>Individual differences in motor noise and adaptation rate are optimally related.</article-title> <source><italic>Eneuro</italic></source> <volume>5</volume>: <comment>ENEURO.0170-18.2018</comment>. <pub-id pub-id-type="doi">10.1523/ENEURO.0170-18.2018</pub-id> <pub-id pub-id-type="pmid">30073197</pub-id></citation></ref>
<ref id="B53"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Villacorta</surname> <given-names>V. M.</given-names></name> <name><surname>Perkell</surname> <given-names>J. S.</given-names></name> <name><surname>Guenther</surname> <given-names>F. H.</given-names></name></person-group> (<year>2007</year>). <article-title>Sensorimotor adaptation to feedback perturbations of vowel acoustics and its relation to perception.</article-title> <source><italic>J. Acoust. Soc. Am.</italic></source> <volume>122</volume> <fpage>2306</fpage>&#x2013;<lpage>2319</lpage>. <pub-id pub-id-type="doi">10.1121/1.2773966</pub-id></citation></ref>
<ref id="B54"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wilcox</surname> <given-names>R. R.</given-names></name> <name><surname>Erceg-Hurn</surname> <given-names>D. M.</given-names></name></person-group> (<year>2012</year>). <article-title>Comparing two dependent groups via quantiles.</article-title> <source><italic>J. Appl. Stat.</italic></source> <volume>39</volume> <fpage>2655</fpage>&#x2013;<lpage>2664</lpage>. <pub-id pub-id-type="doi">10.1080/02664763.2012.724665</pub-id></citation></ref>
<ref id="B55"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wong</surname> <given-names>J.</given-names></name> <name><surname>Wilson</surname> <given-names>E. T.</given-names></name> <name><surname>Malfait</surname> <given-names>N.</given-names></name> <name><surname>Gribble</surname> <given-names>P. L.</given-names></name></person-group> (<year>2009</year>). <article-title>The influence of visual perturbations on the neural control of limb stiffness.</article-title> <source><italic>J. Neurophysiol.</italic></source> <volume>101</volume> <fpage>246</fpage>&#x2013;<lpage>257</lpage>. <pub-id pub-id-type="doi">10.1152/jn.90371.2008</pub-id> <pub-id pub-id-type="pmid">18667545</pub-id></citation></ref>
<ref id="B56"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>H. G.</given-names></name> <name><surname>Miyamoto</surname> <given-names>Y. R.</given-names></name> <name><surname>Castro</surname> <given-names>L. N. G.</given-names></name> <name><surname>&#x00D6;lveczky</surname> <given-names>B. P.</given-names></name> <name><surname>Smith</surname> <given-names>M. A.</given-names></name></person-group> (<year>2014</year>). <article-title>Temporal structure of motor variability is dynamically regulated and predicts motor learning ability.</article-title> <source><italic>Nat. Neurosci.</italic></source> <volume>17</volume> <fpage>312</fpage>&#x2013;<lpage>321</lpage>. <pub-id pub-id-type="doi">10.1038/nn.3616</pub-id> <pub-id pub-id-type="pmid">24413700</pub-id></citation></ref>
</ref-list>
<fn-group>
<fn id="footnote1">
<label>1</label>
<p>Audapter settings for the present study were as follows: sampling rate 48,000 Hz, downsampling factor 3, nDelay factor 3, linear prediction model order 17 for male participants and 15 for female participants. The total feedback loop latency of the specific hardware and software setup is 11.37 milliseconds (<xref ref-type="bibr" rid="B25">Kim et al., 2020b</xref>). Given that Audapter detects vowel onsets and offsets based on a short-time root mean square (RMS) intensity threshold, we determined the optimal RMS threshold for vowel detection for each individual participant based on visual inspection of the RMS intensity contours of the last five trials from the practice/familiarization session.</p></fn>
<fn id="footnote2">
<label>2</label>
<p>The conversion formula between cents and Hz is: <inline-formula><mml:math id="INEQ21"><mml:mrow><mml:mpadded width="+3.3pt"><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>e</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>n</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>t</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>s</mml:mi></mml:mrow></mml:msub></mml:mpadded><mml:mo rspace="5.8pt">=</mml:mo><mml:mrow><mml:mpadded width="+3.3pt"><mml:mn>1200</mml:mn></mml:mpadded><mml:mo rspace="5.8pt">&#x00D7;</mml:mo><mml:mrow><mml:msub><mml:mi>log</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mfrac><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>H</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>z</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mi>R</mml:mi><mml:mrow><mml:mi>H</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>z</mml:mi></mml:mrow></mml:msub></mml:mfrac><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mrow></mml:mrow></mml:math></inline-formula>, where <italic>R</italic><sub><italic>Hz</italic></sub> is a reference frequency. 100 cents = 1 semitone. For the perturbation in the <italic>Adaptation</italic> part, <italic>F</italic><sub><italic>cents</italic></sub>=250, <italic>R</italic><sub><italic>Hz</italic></sub>=<italic>F<sup>c</sup></italic>, and <italic>F</italic><sub><italic>Hz</italic></sub>=<italic>F<sup>fb</sup></italic>. A 250 cents upshift approximately equals a 15.5% increase in Hz.</p></fn>
</fn-group>
</back>
</article>