You're viewing our updated article page. If you need more time to adjust, you can return to the old layout.

MINI REVIEW article

Front. Neurosci., 02 March 2023

Sec. Auditory Cognitive Neuroscience

Volume 17 - 2023 | https://doi.org/10.3389/fnins.2023.1146298

Studying rhythm processing in speech through the lens of auditory-motor synchronization

  • 1. School of Psychology, Beijing Sport University, Beijing, China

  • 2. Laboratory of Sports Stress and Adaptation of General Administration of Sport, Beijing, China

  • 3. Center for the Cognitive Science of Language, Beijing Language and Culture University, Beijing, China

Article metrics

View details

1

Citations

3,5k

Views

1,1k

Downloads

Abstract

Continuous speech is organized into a hierarchy of rhythms. Accurate processing of this rhythmic hierarchy through the interactions of auditory and motor systems is fundamental to speech perception and production. In this mini-review, we aim to evaluate the implementation of behavioral auditory-motor synchronization paradigms when studying rhythm processing in speech. First, we present an overview of the classic finger-tapping paradigm and its application in revealing differences in auditory-motor synchronization between the typical and clinical populations. Next, we highlight key findings on rhythm hierarchy processing in speech and non-speech stimuli from finger-tapping studies. Following this, we discuss the potential caveats of the finger-tapping paradigm and propose the speech-speech synchronization (SSS) task as a promising tool for future studies. Overall, we seek to raise interest in developing new methods to shed light on the neural mechanisms of speech processing.

1. Introduction

Rhythm is a fundamental feature of human speech (Poeppel and Assaneo, 2020). As speech unfolds in time, it is organized into a hierarchy of quasi-rhythmic components along multiple timescales, such as phonemes, syllables, and prosody (Gross et al., 2013). Accurate processing of these hierarchical rhythmic structures lays the foundation of verbal communication.

Researchers have been tackling how the brain processes rhythms in speech for decades. The classic motor theory of speech perception put forward the interaction between auditory and motor systems as a promising solution (Liberman et al., 1967; Liberman and Mattingly, 1985; Galantucci et al., 2006). Accumulating empirical evidence has supported and extended this idea by showing bidirectional auditory-motor interactions in speech rhythm processing. Specifically, motor cortices track speech rhythms during passive listening (e.g., Wilson et al., 2004; Keitel et al., 2018). Meanwhile, speech production also recruits and shapes hearing (e.g., Houde and Jordan, 1998; Assaneo et al., 2021).

However, a critical question remains: How does the auditory-motor interaction work at different rhythmic structure levels? The few studies that addressed this issue revealed inconsistent findings. For example, some showed that motor areas could only track faster and lower-level speech rhythms, i.e., at the syllable level around 4 Hz (Ding et al., 2016; Sheng et al., 2019). Another study reported motor tracking of slower, higher-level rhythm, i.e., at the phrasal level (0.6–1.3 Hz) (Keitel et al., 2018). Moreover, all these studies focused on how auditory rhythms entrain the motor system. Evidence is lacking in the other direction: whether producing different speech rhythms affects hearing differently remains unknown. Therefore, it is time to seek new approaches to solving this question, especially those with behavioral tasks that require rhythm production.

Beyond the realm of speech processing, recent years have witnessed significant progress regarding the role of auditory-motor interaction in general rhythm processing. Many studies exploited a behavioral phenomenon known as auditory-motor synchronization, where participants synchronize their movements to external rhythms. Auditory-motor synchronization can come in many forms of movements, such as cycling limb movements and whole-body dancing (e.g., Chen H.-Y. et al., 2006; Toiviainen et al., 2010; Witek et al., 2017). The most commonly used paradigm is finger tapping, which has a long history (e.g., Fraisse, 1966; Michon, 1967; Keele et al., 1985) and remains popular to this date (for thorough reviews, see Repp, 2005; Repp and Su, 2013). The synchrony performance is assessed by the timing difference between the taps and sound onsets or the variability of inter-tapping time intervals (Repp and Su, 2013). By explicitly translating the rhythm processing results in the brain to behavioral outputs, finger-tapping tasks provide a critical window to the underlying mechanisms.

Converging evidence has demonstrated that auditory-motor synchronization not only reflects basic beat-level perception and action timing (Cannon and Patel, 2021) but also drives more complex forms of rhythm perception, including those involving hierarchical rhythms (Iversen and Balasubramaniam, 2016). Recently, interest has grown in using speech stimuli (Lidji et al., 2011; Falk and Dalla Bella, 2016; Falk et al., 2017; Rathcke et al., 2021; Kliger Amrani and Zion Golumbic, 2022; Wei et al., 2022). Similar to the findings with non-speech stimuli, participants can accurately translate perceived rhythms in the speech into finger-tapping movements. Nevertheless, auditory-motor synchronization paradigms have so far inspired relatively fewer studies on speech rhythms.

This mini-review aims to evaluate the potential implementation of auditory-motor synchronization paradigms in understanding the processing of hierarchical rhythms in speech. First, we summarize key findings relating finger-tapping performance to general speech/language skills. Second, we establish the feasibility of using finger-tapping performance to study the processing of hierarchical rhythms by reviewing studies using non-speech stimuli. Third, we describe current finger-tapping research exploring speech processing. Finally, we address the limitations of the commonly used behavioral paradigms and present a promising new tool.

2. Relating finger-tapping performance to speech processing

While fewer studies used finger-tapping tasks to address rhythm processing in speech, deficits in these tasks have repeatedly been demonstrated in clinical populations (for a thorough review, see Ladányi et al., 2020).

For example, poor finger-tapping performance is found in children with atypical speech/language development, including specific language impairment (Corriveau and Goswami, 2009; Zelaznik and Goffman, 2010; Cumming et al., 2015), speech production disorders (i.e., stuttering) (Falk et al., 2015; Redle et al., 2015), and reading disorders (i.e., dyslexia) (Wolff, 2002; Thomson and Goswami, 2008). Meanwhile, in typically developed children, finger-tapping performance could also predict their speech/language skills, with better synchronizers scoring higher in (pre-)literacy tests and presenting more precise neural encoding of acoustic temporal information (Woodruff Carr et al., 2014; Bonacina et al., 2021; Kertész and Honbolygó, 2021). Finger-tapping deficits have also been found in children and adults who stutter (Smits-Bandstra et al., 2006; Falk et al., 2015).

These results add to the growing literature on the importance of rhythm processing in speech (Hickok et al., 2011; Ladányi et al., 2020). Moreover, the linkage between finger-tapping performance and speech-related deficits indicates a common underlying mechanism, laying the foundation for finger-tapping tasks in speech studies.

3. Studying processing of hierarchical rhythms with finger-tapping to non-speech stimuli

Most finger-tapping studies have featured auditory-motor synchronization at the basic rhythmic (or “beat”) level. Specifically, the most common stimuli were isochronous rhythms (e.g., Aschersleben, 2002; Krause et al., 2010), and participants tapped their fingers to each pulse of the auditory rhythm. Fewer studies addressed rhythms with hierarchical structures, even less with speech rhythms. In this section, we focused on studies using non-speech stimuli and addressed their implications for speech studies.

3.1. Modulating tapping performance by rhythmic structure

One approach to demonstrate auditory-motor interaction in processing higher levels of rhythms is to assess whether varying the rhythmic structure impacts the synchronization performance. When using non-speech stimuli, hierarchical rhythms could be easily introduced by manipulating the acoustic features of the sound elements. Combining with brain imaging techniques, researchers have examined the neural basis of auditory-motor interactions and gained insights into related cognitions (Chen J. L. et al., 2006; Zatorre et al., 2007; Chen J. L. et al., 2008; Kung et al., 2013).

For example, Chen J. L. et al. (2006) introduced hierarchical rhythms to an isochronous tone sequence by placing an intensity accent on every third tone. The attenuation between the unaccented and accented tones was varied, creating different levels of rhythmic saliency. The participants were instructed to synchronize their tapping to each tone, that is, at the basic rhythmic level. Auditory and dorsal premotor cortices showed BOLD responses and functional connectivity covaried with the saliency change. Interestingly, as the saliency increased, accented tones evoked significantly longer taps compared to unaccented tones. Meanwhile, no such difference was found in control trials where the rhythmic structure was compromised.

The rhythm complexity of the stimuli was manipulated in another series of studies (Chen J. L. et al., 2008; Kung et al., 2013). Results showed that accurate timing of tapping (i.e., inter-tap intervals and time asynchrony between tap and sound onsets) depended on the successful resolution of the rhythmic structure and was subjected to musical training experience (Chen J. L. et al., 2008; Kung et al., 2013). The fact that tapping at the basic rhythmic level is affected by changes in the perceived higher-level rhythms indicates a flexible top-down mechanism in coordinating motor outputs with external sounds.

3.2. Interfering rhythm processing by active finger tapping

Since the auditory-motor interaction is bidirectional, another approach is to demonstrate how actively synchronizing movement to external rhythms (as opposed to passive listening) modifies perception. A growing body of evidence has illustrated that tracking auditory rhythms with finger tapping promotes accurate and robust rhythm perception (Morillon et al., 2014; Dalla Bella et al., 2017; Morillon and Baillet, 2017). However, the evidence is sparse concerning higher rhythmic levels.

A recent study on time perception reported direct comparisons between auditory-motor interaction at different rhythmic levels (Hammerschmidt and Wöllner, 2020). In this study, participants were presented with hierarchically structured musical rhythm patterns. They were instructed to tap their fingers to three different hierarchical levels and subsequently estimated the duration of the stimulus. The results showed that tapping to the highest level led to the shortest time estimation. Moreover, participants who tapped more consistently and accurately reported shorter time estimation. Extending these findings to other aspects of temporal processing will shed light on the auditory-motor interaction as the brain interprets speeches unfolding over time (Iversen and Balasubramaniam, 2016).

3.3. Imagining higher-level rhythms: Finger tapping in the absence of external sounds

The finger-tapping task externalizes internal rhythms, with or without an external pacemaker. Constructing imagined rhythm requires the engagement of higher-order brain function in which auditory mental imagery is formed and accurately organized (Lu et al., 2019, 2021). Thus, studying finger-tapping to imagined rhythms could shed light on the top-down mechanisms in encoding and maintaining the hierarchical rhythms.

Motor synchronization to imagined rhythms has been studied using the synchronization-continuation task (e.g., Peters, 1989; Semjen et al., 2000). In this variation of the finger-tapping task, participants first tapped to an external rhythm in the synchronization phase, then reproduced the rhythm without sounds in the continuation phase.

A recent study by Cheng et al. (2022) examined auditory-motor interaction during the imagination of hierarchical rhythms. In each trial, participants first listened to a rhythm containing strong and weak beats (i.e., physical meter condition), then mentally imposed the rhythmic structure onto unaccented sounds (i.e., imagery meter condition). Finally, they reproduced the imagined meter with strong and weak finger taps when the sound stopped. With simultaneous EEG recording, the researchers showed that both auditory and motor neural components tracked the imagined rhythm at the beat and meter rates, with strong bidirectional information flows between auditory and motor systems. In this study, the finger-tapping task mainly served as a verification of correct imagination, thus lacking analyses of neural activities during finger-tapping.

So far, empirical evidence has demonstrated that humans can tap into different levels of rhythm that have differential impacts on perception. These findings indicate finger-tapping tasks as a feasible approach to probing auditory-motor interactions in processing hierarchical rhythms. However, whether this conclusion could be generalized to speech processing requires careful examination, which we addressed in the next section.

4. Studying processing of hierarchical rhythms with finger-tapping to speech stimuli

Direct evidence on whether humans can tap into rhythms in speech has been lacking until recently. In their study, Rathcke et al. (2021) showed that finger-tapping is entrained by natural speech. Moreover, participants were more likely to tap into metrically strong syllables, suggesting that tapping is sensitive to the hierarchical rhythm structure (Rathcke et al., 2021).

However, compared with non-speech stimuli, manipulating rhythmic structure in speech is complicated since speech rhythm is shaped by interacting acoustic and linguistic factors (e.g., Fiveash et al., 2021). Therefore, more elaborate designs are needed.

For example, The English language typically placed the stress at the right edge of each syntactic phrase (e.g., “the BOY that HELPED the GIRL got an “A” on the TEST,” stressed syllables were capitalized). In this way, the acoustic rhythms (also known as meter) aligned with the linguistic/syntactical structure (Hilton and Goldwater, 2021). Hilton and Goldwater (2021) took advantage of this feature and examined the effect of meter-syntax alignment on finger-tapping performance. In each trial, participants were presented with a series of auditory tones followed by an auditory sentence. The sentence stimuli contained monosyllable words, with acoustic features (e.g., duration, intensity, and pitch) normalized among syllables. Thus, there were no acoustic cues to meter in these sentences. Meanwhile, the preceding tones were amplitude modulated so that the strong and weak tones set a metrical context that either aligned or misaligned with the syntactical structure of the sentence. Participants tapped their fingers in time with the strong beat of the tones, then continued the pattern on their own during the sentence presentation. At the end of the trial, participants completed a comprehension task. Misaligning meter and syntax caused more comprehension mistakes and, intriguingly, disrupted auditory-motor synchronization: participants showed larger tapping variability during the sentence presentation. This study adds new evidence to the growing literature on the auditory modulation of motor output. More importantly, it provides behavioral evidence for the hypothesis that auditory-motor interaction optimizes speech comprehension, which may explain our natural tendency to gesture while speaking (Hilton and Goldwater, 2021).

Meanwhile, studies in the linguistic field have demonstrated the facilitating role of auditory-motor interaction in speech perception (Falk and Dalla Bella, 2016; Falk et al., 2017). Researchers composed speech stimuli alternating stressed (i.e., accent) and unstressed syllables, thus forming an acoustic hierarchical structure or meter. Participants were instructed to align (“congruent” condition) or misalign (“incongruent” condition) their finger tapping to the stressed syllables while performing a word change detection task. The results showed that congruent alignment of motor rhythm with the stressed syllables resulted in better detection performance than the incongruent condition and perceptual control conditions without finger tapping (Falk and Dalla Bella, 2016). Further, in the congruent condition, detection performance was predicted by participants’ tapping accuracy to non-verbal cueing stimuli (Falk et al., 2017). These findings were consistent with the notion that motor synchronization contributes to auditory temporal attention (Morillon et al., 2014). Notably, in the studies above, the absolute tapping rates were constant among trials. Whether auditory-motor interaction to different rhythmic levels differs in effects remained unanswered.

5. Studying auditory-motor synchronization in rhythm processing with the speech-to-speech synchronization task

Aside from the fruitful findings, controversy remains: is finger-tapping, a non-verbal behavior, effective in revealing auditory-motor synchronization specific to speech? The auditory-motor synchronization phenomenon has been found in other effectors, such as foot, head, and torso movements (Bouwer et al., 2021), with evidence that different effectors synchronized to different levels of rhythmic hierarchy (Burger et al., 2014; Pflug et al., 2019). For example, a recent study demonstrated a discrepancy between the dominant hand and the non-dominant hand: for right-handed participants, faster tapping at a beat level was better with the right hand, and slower tapping at a higher level was better with the left hand (Pflug et al., 2019). Moreover, finger tapping and speech production involve different optimal time scales: while the auditory-motor interaction in speech processing is restricted to around 4.5 Hz (Assaneo and Poeppel, 2018), finger tapping could adapt to a wider range of tempi, with an optimal frequency around 1.5 Hz (Zalta et al., 2020). Therefore, conclusions should be made with caution, considering the possible discrepancies between fingers and speech effectors when studying speech-specific auditory-motor interaction.

The recently developed spontaneous synchronization of speech (also referred to as speech-to-speech synchronization, SSS) task evaluates participants’ ability to synchronize speech production, a verbal behavior, with external speech sounds (Assaneo et al., 2019; Lizcano-Cortés et al., 2022). Thus, the SSS task provides a direct measurement of speech-specific auditory-motor synchronization. In this task, participants synchronize their vocalization to syllable sequences presented at a constant rate (typically at 4.5 Hz) or an accelerating rate (e.g., 4.3 to 4.7 Hz in 0.1 Hz steps). The subject vocalization is recorded, and the phase locking value (PLV) of the reconstructed envelope of the recorded and presented signals is computed to measure the synchronization of speech.

The original (Assaneo et al., 2019) and the follow-up studies (Assaneo et al., 2021; Kern et al., 2021) showed that the SSS task stably classified the general population into two groups: high synchronizers and low synchronizers. Aside from superior speech synchronization, several advantages in speech-related cognitions have been demonstrated for high synchronizers, including better performance when learning pseudo-words (Assaneo et al., 2019; Orpella et al., 2022) and discriminating rate differences beyond the typical optimal range (Kern et al., 2021). High synchronizers also showed stronger motor entrainment of speech perception (Assaneo et al., 2021). In this study, subjects performed a syllable discrimination task after rhythmically producing syllable sequences. In high synchronizers only, the perceptual performance was modulated by the phase of the target syllable, which was determined by the production rhythm. Although fewer studies explore the neural basis underlying the behavioral differences, distinct brain features have been found between high and low synchronizers (Assaneo et al., 2019). Specifically, high synchronizers demonstrated enhanced brain-stimulus synchrony in the frontal area, better synchrony symmetry in the early auditory area, and greater left lateralization in the white matter connection between these two areas.

This drastic dichotomy of low- and high-synchronizers has not been spotted by the vast body of finger-tapping studies. Therefore, the SSS task may be more sensitive to speech-specific rhythm processing. To be noted, by far, the SSS task has only been validated in native English (Assaneo et al., 2019) and German speakers (Assaneo et al., 2021). Further studies are needed to assess the generalization to other language populations.

6. Discussion

Auditory-motor interaction in speech processing has been an active field for decades (Hickok et al., 2011; Poeppel and Assaneo, 2020). Emerging evidence from finger-tapping studies suggests that this interaction exists not only at the basic rhythmic level (Rathcke et al., 2021; Kliger Amrani and Zion Golumbic, 2022) but also at higher levels, subject to variations in both acoustic and syntactical structures (Falk and Dalla Bella, 2016; Falk et al., 2017; Hilton and Goldwater, 2021). Moreover, the auditory-motor interaction is bidirectional: the perception of higher-level rhythms affects motor synchronization to external sounds (Chen J. L. et al., 2006; Chen J. L. et al., 2008; Kung et al., 2013; Hilton and Goldwater, 2021); meanwhile, active motor synchronization also affects the perception of the sounds (Falk and Dalla Bella, 2016; Falk et al., 2017; Assaneo et al., 2021). However, direct evidence is still lacking. More studies using speech stimuli are needed, as well as direct comparisons of the auditory-motor interaction mechanisms at the basic and higher rhythmic levels. Moreover, as mentioned earlier, the newly developed SSS task measures speech-specific auditory-motor interaction using vocalization as the targeted movement. To our knowledge, no studies have yet used SSS tasks in the context of multiple rhythmic levels. Inspired by finger-tapping studies, foreseen future directions include manipulating the rhythmic structures of the syllables sequence and requiring vocalization to a specific level of the hierarchical structure.

Statements

Author contributions

LL and LxL conceived the focus of the review, reviewed the literature, and finalized the manuscript. Both authors approved the final version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (32100866), the International Chinese Language Education Research Program of the Center for Language Education and Cooperation (22YH22B), and the Research Project “The Construction of the Advanced Disciplines in Universities in Beijing”.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  • 1

    Aschersleben G. (2002). Temporal control of movements in sensorimotor synchronization.Brain Cogn.486679. 10.1006/brcg.2001.1304

  • 2

    Assaneo M. F. Poeppel D. (2018). The coupling between auditory and motor cortices is rate-restricted: Evidence for an intrinsic speech-motor rhythm.Sci. Adv.4:eaao3842. 10.1126/sciadv.aao3842

  • 3

    Assaneo M. F. Rimmele J. M. Sanz Perl Y. Poeppel D. (2021). Speaking rhythmically can shape hearing.Nat. Hum. Behav.57182. 10.1038/s41562-020-00962-0

  • 4

    Assaneo M. F. Ripollés P. Orpella J. Lin W. M. de Diego-Balaguer R. Poeppel D. (2019). Spontaneous synchronization to speech reveals neural mechanisms facilitating language learning.Nat. Neurosci.22627632. 10.1038/s41593-019-0353-z

  • 5

    Bonacina S. Huang S. White-Schwoch T. Krizman J. Nicol T. Kraus N. (2021). Rhythm, reading, and sound processing in the brain in preschool children.NPJ Sci. Learn.6:20. 10.1038/s41539-021-00097-5

  • 6

    Bouwer F. L. Nityananda V. Rouse A. A. Ten Cate C. (2021). Rhythmic abilities in humans and non-human animals: A review and recommendations from a methodological perspective.Phil. Trans. R. Soc. B376:20200335. 10.1098/rstb.2020.0335

  • 7

    Burger B. Thompson M. R. Luck G. Saarikallio S. H. Toiviainen P. (2014). Hunting for the beat in the body: On period and phase locking in music-induced movement.Front. Hum. Neurosci.8:903. 10.3389/fnhum.2014.00903

  • 8

    Chen H.-Y. Wing A. M. Pratt D. (2006). The synchronization of lower limb responses with a variable metronome: The effect of biomechanical constraints on timing.Gait Posture23307314. 10.1016/j.gaitpost.2005.04.001

  • 9

    Cannon J. J. Patel A. D. (2021). How beat perception co-opts motor neurophysiology.Trends Cogn. Sci.25137150. 10.1016/j.tics.2020.11.002

  • 10

    Chen J. L. Penhune V. B. Zatorre R. J. (2008). Moving on time: Brain network for auditory-motor synchronization is modulated by rhythm complexity and musical training.J. Cogn. Neurosci.20226239. 10.1162/jocn.2008.20018

  • 11

    Chen J. L. Zatorre R. J. Penhune V. B. (2006). Interactions between auditory and dorsal premotor cortex during synchronization to musical rhythms.Neuroimage3217711781. 10.1016/j.neuroimage.2006.04.207

  • 12

    Cheng T.-H. Z. Creel S. C. Iversen J. R. (2022). How do you feel the rhythm: Dynamic motor-auditory interactions are involved in the imagination of hierarchical timing.J. Neurosci.42500512. 10.1523/JNEUROSCI.1121-21.2021

  • 13

    Corriveau K. H. Goswami U. (2009). Rhythmic motor entrainment in children with speech and language impairments: Tapping to the beat.Cortex45119130. 10.1016/j.cortex.2007.09.008

  • 14

    Cumming R. Wilson A. Leong V. Colling L. J. Goswami U. (2015). Awareness of rhythm patterns in speech and music in children with specific language impairments.Front. Hum. Neurosci.9:672. 10.3389/fnhum.2015.00672

  • 15

    Dalla Bella S. Farrugia N. Benoit C.-E. Begel V. Verga L. Harding E. et al (2017). BAASTA: Battery for the assessment of auditory sensorimotor and timing abilities.Behav. Res.4911281145. 10.3758/s13428-016-0773-6

  • 16

    Ding N. Melloni L. Zhang H. Tian X. Poeppel D. (2016). Cortical tracking of hierarchical linguistic structures in connected speech.Nat. Neurosci.19158164. 10.1038/nn.4186

  • 17

    Falk S. Dalla Bella S. (2016). It is better when expected: Aligning speech and motor rhythms enhances verbal processing.Lang. Cogn. Neurosci.31699708. 10.1080/23273798.2016.1144892

  • 18

    Falk S. Müller T. Dalla Bella S. (2015). Non-verbal sensorimotor timing deficits in children and adolescents who stutter.Front. Psychol.6:847. 10.3389/fpsyg.2015.00847

  • 19

    Falk S. Volpi-Moncorger C. Dalla Bella S. (2017). Auditory-motor rhythms and speech processing in French and German listeners.Front. Psychol.8:395. 10.3389/fpsyg.2017.00395

  • 20

    Fraisse P. (1966). Anticipation of rhythmic stimuli: Speed of establishment and precision of synchronization.Annee. Psychol.661536. 10.3406/psy.1966.27876

  • 21

    Fiveash A. Bedoin N. Gordon R. L. Tillmann B. (2021). Processing rhythm in speech and music: Shared mechanisms and implications for developmental speech and language disorders.Neuropsychology35:771. 10.1037/neu0000766

  • 22

    Galantucci B. Fowler C. A. Turvey M. T. (2006). The motor theory of speech perception reviewed.Psychon. Bull. Rev.13361377. 10.3758/BF03193857

  • 23

    Gross J. Hoogenboom N. Thut G. Schyns P. Panzeri S. Belin P. et al (2013). Speech rhythms and multiplexed oscillatory sensory coding in the human brain.PLoS Biol.11:e1001752. 10.1371/journal.pbio.1001752

  • 24

    Hammerschmidt D. Wöllner C. (2020). Sensorimotor synchronization with higher metrical levels in music shortens perceived time.Music Percept.37263277. 10.1525/mp.2020.37.4.263

  • 25

    Hickok G. Houde J. Rong F. (2011). Sensorimotor integration in speech processing: Computational basis and neural organization.Neuron69407422. 10.1016/j.neuron.2011.01.019

  • 26

    Hilton C. B. Goldwater M. B. (2021). Linguistic syncopation: Meter-syntax alignment affects sentence comprehension and sensorimotor synchronization.Cognition217:104880. 10.1016/j.cognition.2021.104880

  • 27

    Houde J. F. Jordan M. I. (1998). Sensorimotor adaptation in speech production.Science27912131216. 10.1126/science.279.5354.1213

  • 28

    Iversen J. R. Balasubramaniam R. (2016). Synchronization and temporal processing.Curr. Opin. Behav. Sci.8175180. 10.1016/j.cobeha.2016.02.027

  • 29

    Keitel A. Gross J. Kayser C. (2018). Perceptually relevant speech tracking in auditory and motor cortex reflects distinct linguistic features.PLoS Biol.16:e2004473. 10.1371/journal.pbio.2004473

  • 30

    Keele S. W. Pokorny R. A. Corcos D. M. Ivry R. (1985). Do perception and motor production share common timing mechanisms: A correlational analysis.Acta Psychol.60173191. 10.1016/0001-6918(85)90054-X

  • 31

    Kern P. Assaneo M. F. Endres D. Poeppel D. Rimmele J. M. (2021). Preferred auditory temporal processing regimes and auditory-motor synchronization.Psychon. Bull. Rev.2818601873. 10.3758/s13423-021-01933-w

  • 32

    Kertész C. Honbolygó F. (2021). Tapping to music predicts literacy skills of first-grade children.Front. Psychol.12:4412. 10.3389/fpsyg.2021.741540

  • 33

    Kliger Amrani A. Zion Golumbic E. (2022). Memory-paced tapping to auditory rhythms: Effects of rate, Speech, and Motor Engagement.J. Speech Lang. Hear. Res.65923939. 10.1044/2021_JSLHR-21-00406

  • 34

    Krause V. Schnitzler A. Pollok B. (2010). Functional network interactions during sensorimotor synchronization in musicians and non-musicians.Neuroimage52245251. 10.1016/j.neuroimage.2010.03.081

  • 35

    Kung S.-J. Chen J. L. Zatorre R. J. Penhune V. B. (2013). Interacting cortical and basal ganglia networks underlying finding and tapping to the musical beat.J. Cogn. Neurosci.25401420. 10.1162/jocn_a_00325

  • 36

    Ladányi E. Persici V. Fiveash A. Tillmann B. Gordon R. L. (2020). Is atypical rhythm a risk factor for developmental speech and language disorders?Wiley Interdiscip. Rev. Cogn. Sci.11e1528. 10.1002/wcs.1528

  • 37

    Liberman A. M. Cooper F. S. Shankweiler D. P. Studdert-Kennedy M. (1967). Perception of the speech code.Psychol. Rev.74:431. 10.1037/h0020279

  • 38

    Liberman A. M. Mattingly I. G. (1985). The motor theory of speech perception revised.Cognition21136. 10.1016/0010-0277(85)90021-6

  • 39

    Lidji P. Palmer C. Peretz I. Morningstar M. (2011). Listeners feel the beat: Entrainment to English and French speech rhythms.Psychon. Bull. Rev.1810351041. 10.3758/s13423-011-0163-0

  • 40

    Lizcano-Cortés F. Gómez-Varela I. Mares C. Wallisch P. Orpella J. Poeppel D. et al (2022). Speech-to-Speech Synchronization protocol to classify human participants as high or low auditory-motor synchronizers.STAR Protoc.3:101248. 10.1016/j.xpro.2022.101248

  • 41

    Lu L. Sheng J. Liu Z. Gao J.-H. (2021). Neural representations of imagined speech revealed by frequency-tagged magnetoencephalography responses.Neuroimage229:117724. 10.1016/j.neuroimage.2021.117724

  • 42

    Lu L. Wang Q. Sheng J. Liu Z. Qin L. Li L. et al (2019). Neural tracking of speech mental imagery during rhythmic inner counting.Elife8:e48971. 10.7554/eLife.48971

  • 43

    Michon J. A. (1967). Timing in temporal tracking.Assen: van Gorcum.

  • 44

    Morillon B. Baillet S. (2017). Motor origin of temporal predictions in auditory attention.Proc. Natl. Acad. Sci. U.S.A.114E8913E8921. 10.1073/pnas.1705373114

  • 45

    Morillon B. Schroeder C. E. Wyart V. (2014). Motor contributions to the temporal precision of auditory attention.Nat. Commun.5:5255. 10.1038/ncomms6255

  • 46

    Orpella J. Assaneo M. F. Ripollés P. Noejovich L. López-Barroso D. Diego-Balaguer R. D. et al (2022). Differential activation of a frontoparietal network explains population-level differences in statistical learning from speech.PLoS Biol.20:e3001712. 10.1371/journal.pbio.3001712

  • 47

    Peters M. (1989). The relationship between variability of intertap intervals and interval duration.Psychol. Res.513842. 10.1007/BF00309274

  • 48

    Pflug A. Gompf F. Muthuraman M. Groppa S. Kell C. A. (2019). Differential contributions of the two human cerebral hemispheres to action timing.Elife8:e48404. 10.7554/eLife.48404

  • 49

    Poeppel D. Assaneo M. F. (2020). Speech rhythms and their neural foundations.Nat. Rev. Neurosci.21322334. 10.1038/s41583-020-0304-4

  • 50

    Rathcke T. Lin C.-Y. Falk S. Bella S. D. (2021). Tapping into linguistic rhythm.Lab. Phonol.12111. 10.5334/labphon.248

  • 51

    Redle E. Vannest J. Maloney T. Tsevat R. K. Eikenberry S. Lewis B. et al (2015). Functional MRI evidence for fine motor praxis dysfunction in children with persistent speech disorders.Brain Res.15974756. 10.1016/j.brainres.2014.11.047

  • 52

    Repp B. H. (2005). Sensorimotor synchronization: A review of the tapping literature.Psychon. Bull. Rev.12969992. 10.3758/BF03206433

  • 53

    Repp B. H. Su Y.-H. (2013). Sensorimotor synchronization: A review of recent research (2006–2012).Psychon. Bull. Rev.20403452. 10.3758/s13423-012-0371-2

  • 54

    Semjen A. Schulze H.-H. Vorberg D. (2000). Timing precision in continuation and synchronization tapping. Psychol. Res.63, 137147. 10.1007/PL00008172

  • 55

    Sheng J. Zheng L. Lyu B. Cen Z. Qin L. Tan L. H. et al (2019). The cortical maps of hierarchical linguistic structures during speech perception.Cereb. Cortex2932323240. 10.1093/cercor/bhy191

  • 56

    Smits-Bandstra S. Luc F. Saint-Cyr J. A. (2006). Speech and nonspeech sequence skill learning in adults who stutter.J. Fluen. Disord.31116136. 10.1016/j.jfludis.2006.04.003

  • 57

    Thomson J. M. Goswami U. (2008). Rhythmic processing in children with developmental dyslexia: Auditory and motor rhythms link to reading and spelling.J. Physiol. Paris102120129. 10.1016/j.jphysparis.2008.03.007

  • 58

    Toiviainen P. Luck G. Thompson M. R. (2010). Embodied meter: Hierarchical eigenmodes in music-induced movement. Music Percept.28, 5970. 10.1525/mp.2010.28.1.59

  • 59

    Wei Y. Hancock R. Mozeiko J. Large E. W. (2022). The relationship between entrainment dynamics and reading fluency assessed by sensorimotor perturbation.Exp. Brain Res.24017751790. 10.1007/s00221-022-06369-9

  • 60

    Wilson S. M. Saygin A. P. Sereno M. I. Iacoboni M. (2004). Listening to speech activates motor areas involved in speech production.Nat. Neurosci.7701702. 10.1038/nn1263

  • 61

    Witek M. A. Popescu T. Clarke E. F. Hansen M. Konvalinka I. Kringelbach M. L. et al (2017). Syncopation affects free body-movement in musical groove. Exp. Brain Res. 235, 9951005. 10.1007/s00221-016-4855-6

  • 62

    Wolff P. H. (2002). Timing precision and rhythm in developmental dyslexia.Read. Writ.15179206. 10.1023/A:1013880723925

  • 63

    Woodruff Carr K. White-Schwoch T. Tierney A. T. Strait D. L. Kraus N. (2014). Beat synchronization predicts neural speech encoding and reading readiness in preschoolers.Proc. Natl. Acad. Sci. U.S.A.1111455914564. 10.1073/pnas.1406219111

  • 64

    Zalta A. Petkoski S. Morillon B. (2020). Natural rhythms of periodic temporal attention.Nat. Commun.11:1051. 10.1038/s41467-020-14888-8

  • 65

    Zatorre R. J. Chen J. L. Penhune V. B. (2007). When the brain plays music: Auditory-motor interactions in music perception and production.Nat. Rev. Neurosci.8547558. 10.1038/nrn2152

  • 66

    Zelaznik H. N. Goffman L. (2010). Generalized motor abilities and timing behavior in children with specific language impairment.J. Speech Lang. Hear. Res.53383393. 10.1044/1092-4388(2009/08-0204)

Summary

Keywords

auditory-motor synchronization, rhythm, speech, finger-tapping task, speech-to-speech synchronization task

Citation

Luo L and Lu L (2023) Studying rhythm processing in speech through the lens of auditory-motor synchronization. Front. Neurosci. 17:1146298. doi: 10.3389/fnins.2023.1146298

Received

17 January 2023

Accepted

20 February 2023

Published

02 March 2023

Volume

17 - 2023

Edited by

Dan Zhang, Tsinghua University, China

Reviewed by

Jiawei Li, Freie Universität Berlin, Germany

Updates

Copyright

*Correspondence: Lingxi Lu,

This article was submitted to Auditory Cognitive Neuroscience, a section of the journal Frontiers in Neuroscience

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics