Your new experience awaits. Try the new design now and help us make it even better

HYPOTHESIS AND THEORY article

Front. Cognit., 06 February 2026

Sec. Perception

Volume 4 - 2025 | https://doi.org/10.3389/fcogn.2025.1689600

This article is part of the Research TopicNeurocognitive Bases of Music ReadingView all 8 articles

Bi-temporal processing in music notation reading: a theory linking prediction, memory, and automaticity

  • Melbourne Conservatorium of Music, The University of Melbourne, Melbourne, VIC, Australia

Reading music notation requires musicians to extract and interpret visual information in real time while simultaneously anticipating future performance actions. This dual engagement, in which one acts in the present while processing material to be performed in the future, suggests that music reading relies on a bi-temporal cognitive architecture. Grounded in this premise, this theoretical paper develops a model that integrates Hebbian learning and automaticity as core mechanisms supporting the simultaneous perceptual and anticipatory demands of notation-based music performance. A systematic review of neuroimaging studies involving music-reading tasks was conducted to evaluate current evidence on the neural correlates of notation processing. The results of the review showed that music reading engaged distributed cortical and subcortical networks, including regions commonly implicated in text reading, and recruited auditory-motor integration systems essential for music performance. However, most studies isolated single parameters of notation (e.g., pitch identification), thereby limiting ecological validity and constraining interpretations of how musicians process in real-world contexts that require concurrent multi-parameter integration. Complementary research on cognitive prediction, sensorimotor coupling, and perceptual-motor learning demonstrates that musicians employ a dual-pathway system of immediate perception and forward prediction, shaped by Hebbian synaptic strengthening and the development of automaticity through repeated procedural engagement. Synthesizing these findings, this article proposes a bi-temporal cognitive model of music-notation processing that accounts for dynamic interplay between associative learning, predictive processing, and automated motor execution. The implications of this model for cognitive theory and music pedagogy are discussed, with recommendations for empirical approaches to test the bi-temporal framework and advance understanding of real-time cognitive coordination in music performance.

1 Introduction

Reading music notation is a fundamental skill for professional musicians who work in orchestral, ensemble, educational, and sessional performance contexts. For professional and/or highly skilled amateur musicians who perform using music notation, musical outcomes arise from the simultaneous processing of auditory and visual information, as well as motoric actions (Bigand et al., 2000; Gaser and Gottfried Schlaug, 2003; Ho et al., 2003; Zimmerman and Lahav, 2012). The notion that musical performance encompasses a range of different functions, perception modalities, and physiological outputs is extensively supported in the literature. Research focused on music in psychological, neurological, and physiological contexts points to significant complexities involved with cognition (Schön and Besson, 2002; Gunter et al., 2003; Stewart, 2005a,b; Sluming et al., 2007), brain plasticity (Stewart et al., 2003a; Stewart, 2005a), memory (Simoens and Tervaniemi, 2013), emotion (Jäncke, 2012; Koshimori, 2018; Schubert, 2013), spatial mapping (Stewart, 2005a), and somatosensory perceptions (Brancucci and San Martini, 1999, 2003; Meister et al., 2004; Brancucci et al., 2005; Wong and Gauthier, 2010b).

1.1 Music-notation reading as a cognitive subdomain

In this article, reading music notation is positioned as a specific subdomain of musical cognition operating at the interface of perception and performance. For analytic purposes, the component processes of reading music are classified as follows: (a) visual-symbolic decoding, in which elements found in notated scores are mapped onto abstract pitch, rhythmic, and structural representations; (b) audiative-predictive mapping, where notation is transformed into sonic representations, perceived as mental imagery for sounds yet to be performed; and (c) sensorimotor execution, in which predicted events are realized through motor functions. In this sense, we can see that reading and interpreting music notation is not an isolated skill but rather a notation-based entry point into a larger performance system that coordinates perceptual, predictive, and motor processes within a bi-temporal architecture. This frame of reference allows for both present-focused sensorimotor execution and future-oriented audiative–predictive processing, both anchored in ongoing visual–symbolic decoding.

To explore the cognitive processes responsible for interpreting music notation, we must first acknowledge the multivariate, simultaneous events involved in performing and reconcile them with current viewpoints on multitasking. It is now well-understood that the human brain is unable to consciously make decisions about multiple events simultaneously or hold more than one novel thought at a time (Gazzaley and Rosen, 2016; Hari, 2022). How, then, do we account for the multiple, concurrent, and bi-temporal processing of music notation, aural stimuli, motor action responses, anticipatory sensorimotor planning, and the ensuing musical outcomes? Even just the act of performing motor functions requires continuous updating by sensory systems, using information derived from external stimuli (Kandel et al., 2014).

In consideration of possible explanations for these phenomena, this paper explores the roles that Hebbian learning and automaticity play in reading music notation and hypothesizes a theory of cognitive processing that incorporates a neural basis for learning and memory formation and the ability to operate in two time zones. The theoretical framework presented in this article is based on the premise that cognition operates in different temporal horizons in which past information is accessed while future actions are anticipated (Schacter et al., 2007, 2017; Tulving, 2002). Following this, the argument is presented that musicians with high-level skills in reading music notation process information in a dual-stimulus paradigm that encompasses both visual and aural categories, resulting in a unified, cohesive musical output.

This paper commences with a summary of frameworks, theories, and concepts that underpin the theory presented for consideration. The second section reviews the literature and evaluates previously proposed hypotheses, theories, and current perspectives related to the neural bases of reading musical notation. Included in this section is a systematic review of brain imaging studies that incorporate music notation reading. A discussion section follows in which the theory central to this paper is presented in full, along with suggestions for testing within the framework of current knowledge. Concluding remarks detail possible implications of integrating a bi-temporal component in current understandings of cognitive processing when reading musical notation. Further investigations in this area could offer a critical contribution to understanding neuronal pathway operations during music reading, with potential implications for music education as well as other domains of learning.

1.2 Aims and research questions

The central aim of this study is to present a theory on cognitive mechanisms underlying the processing demands of reading musical notation. Emphasis is given to Hebbian learning and automaticity to support the notion of simultaneous perception and prediction in performance. This inquiry responds to persistent limitations in existing empirical work, which has largely examined isolated parameters of notation such as basic pitch identification or simple rhythmic units. This mode of inquiry, which individuates the various task parameters of playing music from notation, contrasts with the reality experienced by musicians when playing from notation, which necessitates the concurrent processing of numerous visual stimuli (Bigand et al., 2000; Gaser and Gottfried Schlaug, 2003; Ho et al., 2003; Zimmerman and Lahav, 2012). The outcome of synthesizing neuroimaging evidence with broader cognitive research on prediction, cognitive load theory (Sweller, 2011), sensorimotor coupling, and perceptual-motor learning (Schön and Besson, 2002; Gunter et al., 2003; Stewart, 2005a,b; Sluming et al., 2007), is a comprehensive bi-temporal model of music notation processing.

To guide the inquiry and subsequent development of a bi-temporal model of music notation reading, the following research questions were posited:

• What neural systems are consistently implicated in real-time notation reading, and how do these compare to those recruited for text processing and auditory-motor integration?

• How do Hebbian learning mechanisms and automaticity contribute to musicians' capacity to engage in simultaneous present-focused perception and future-oriented prediction?

• To what extent does current literature support or constrain a theoretical account of bi-temporal cognitive architecture in music performance?

• What methodological approaches are required to empirically test this proposed model and more accurately capture dual-pathway perceptual-predictive processing in ecologically valid contexts?

2 The neural architecture of music notation reading

To clarify the central thesis of this paper, reading music is classified into three processes: visual-symbolic decoding, audiative-predictive mapping, and sensorimotor execution. Each of these components aligns with specific neural structures. The first component, the visual-symbolic decoding process, involves the occipito-temporal visual system extracting the orthographic features of notation (e.g., notes, clefs, accidentals, etc.) and mapping them onto pitch-rhythm representations. This process depends on ventral visual pathways, including areas that show expertise-related tuning (Bouhali et al., 2017; Proverbio et al., 2024). The second component, audiative-predictive mapping, is carried out by transforming visual information into internally generated auditory imagery, or audiation (Gordon, 1999, 2011), along with predictive models of upcoming sound events. This engages predictive planning mechanisms supported by fronto-parietal systems, including working memory (WM) buffers and the dorsal attention network (DAN), aiding future-focused planning (Palmer and van de Sande, 1993; Wolff et al., 2022). The final component, sensorimotor execution, involves translating predicted musical events into motor programs via premotor, motor, cerebellar, and basal ganglia circuits. These same systems facilitate proceduralization and automaticity (Ashby and Crossley, 2012; Bangert et al., 2006). Central to this paper is the idea that these components operate on two time scales: one governs the real-time execution of previously read material, and the other operates in a predictive mode, where visual information is processed and mapped onto various aspects of music reading.

2.1 The musician's process: performance in a bi-temporal paradigm

When a musician has practiced extensively to prepare for a performance, successful musical outcomes are often attributed to muscle memory or to the performer being well-rehearsed. This a priori assessment provides a satisfactory explanation of how musicians can carry out numerous actions simultaneously while thinking several steps ahead—that is, until we inquire into the cognitive processing necessary for musicians to sightread, i.e., to play from sheet music without any rehearsal or preparation whatsoever. Muscle memory through repetitive practice can no longer be considered the underlying reason, at least not without some interrogation into the different conditions of performance, such as the level of difficulty, familiarity with the patterns embedded within the music, and the overall predictability of the notation being performed.

Mastering a musical instrument involves the acquisition of numerous skills, some of which are refined to the point of automaticity. Research indicates that varying degrees of automaticity are evident in domains such as language acquisition and mathematics (Jeon and Friederici, 2015). However, because music performance has multiple components, there is a distinction between the operations of music and other activities, with potentially numerous parallel memory retrieval systems operating at varying levels of automaticity (Heath, 2025). Additionally, automaticity occurs within the confines of a strict temporal structure as an essential component of music performance (Wakita, 2016). Within the boundaries of time, a high-level professional musician can typically perform repertoire with little need for conscious control over certain aspects of their playing; instead, a series of instinctive, reflexive actions seems to take place, even when performing novel material from music notation.

The element of time is a crucial aspect in music performance. For one, music is an art form grounded in temporality. Secondly, the prospective planning involved in performing from notation at a high level of competency requires a musician to anticipate future actions while executing multiple tasks simultaneously (Palmer and Drake, 1997). This suggests that dual processing may occur across a range of brain regions, operating in two distinct time zones. In a review of the literature (Lewis and Miall, 2003), evidence is presented for separate neural timing systems, in which specific tasks utilize different systems for categorizing temporality. Here, “automatic” actions relied on circuitry within the motor system, whereas time-based actions required conscious control, actualized in the prefrontal and parietal regions (Lewis and Miall, 2003). More recent research highlights a distinction between time in terms of cognitive perception, defined as prospective timing for events that begin in the present and end in the future, and retrospective timing—events that began in the past and are to end in the past or the present (Tsao et al., 2022). Here, we observe a connection to how musicians read music as they read ahead, conceptualizing future actions while performing previously read notation in the present moment.

With evidence that the brain can encode temporal information over spans ranging from 1 s to several minutes (Tsao et al., 2022), the notion of dual processing is further supported by Høffding (2014) in a phenomenological study of performance as experienced by musicians. In this study, which situates phenomenological data illustratively rather than as empirical evidence, one of the participants describes the operation of performing music as “two tracks running” in which “an awareness of what you are doing and an awareness of what you would like to do…” (Høffding, 2014, p. 65). For musicians, dual operations through prospective and retrospective planning, within the constraints of time, are fundamental to performance.

The initial process of acquiring any skill begins with focused attention and engagement, facilitated by executive functioning and working memory (WM). According to Sachs and colleagues, multiple sensorimotor processes associated with performing music are understood to be operationalized by these areas of cognition (Sachs et al., 2017). As a cognitive processing action, executive functioning encompasses decision-making, judgment, evaluation, and planning (Gazzaley and Rosen, 2016). WM is the temporary storage of information that allows for planning, responding, comprehending, and decision-making (Baddeley and Hitch, 1974; George and Coch, 2011; Schulze and Koelsch, 2012). Although music performance via notation requires a broad engagement of different brain areas simultaneously (Lee et al., 2007; Olivers and Eimer, 2010; Sala and Gobet, 2017; Schulze and Koelsch, 2012; Silverman, 2010; Yeşil and Nal, 2017; Yurgil et al., 2020), research suggests that performing multiple tasks simultaneously—such as the array of actions required to play music—presents significant limitations because the human brain can only attend to one conscious, or deliberate, thought at a time (Gazzaley and Rosen, 2016; Olivers and Meeter, 2008). A convergence of literature relating to attention through mechanisms such as working memory and executive functioning, which neurobiologically operate within the dorsolateral prefrontal cortex (DLPFC), has determined that attempting to engage in numerous actions at once drastically depletes the fidelity of focus and concentration (Gazzaley and Rosen, 2016; Wieth and Burns, 2014). In essence, what was once referred to as multitasking is now considered an action that significantly reduces accuracy, leading to cognitive overload, divided attention, and frequent task switching (Gazzaley and Rosen, 2016; Strobach et al., 2018; Wieth and Burns, 2014).

To distinguish the tasks that a musician is required to undertake simultaneously while also in a state of prospective planning, a brief list of actions might include the following: (a) audiation, the process of conceptualizing and hearing music in one's mind when reading musical notation (Gordon, 2011); (b) plotting out technical considerations in advance, such as bowing patterns, finger placements, and articulation; and (c) accurately playing, in real-time, music notation that had been read only moments earlier. If we accept that a musician performs these three discrete tasks simultaneously, we may conclude that there is a set of cognitive processes controlling each task, each sufficiently automated to enable an immediate response to aural and visual stimuli. After the process of working memory and executive function has enabled the musician to learn materials and skills to a sufficient level of fluency, cognitive processing is offloaded to motor-based systems without the involvement of the prefrontal cortex (PFC; Chein and Schneider, 2005; Kelly and Garavan, 2005). Supporting this premise is the theory of automaticity as applied to performing music from notation stimuli, similar to that of reading text.

2.2 Neural architecture: attentional and executive networks

Research in cognitive neuroscience demonstrates that executive function, working memory, and attentional control are not reducible to a single cortical region. Instead, they emerge from coordinated activity across large-scale networks, including the Executive Control Network (ECN), the Dorsal Attention Network (DAN), and the Ventral Attention Network (VAN). During music-notation reading, these networks support shifting between present-focused motor execution and future-focused perceptual prediction, enabling the bi-temporal processing proposed in this model. In this model, executive and attentional networks are not treated as independent causal agents, but as interacting systems that enable or constrain bi-temporal processing. While dorsolateral prefrontal cortex (DLPFC) recruitment has been observed in early stages of learning, fluent performance increasingly relies on distributed network interactions rather than isolated regional activation.

2.3 Automaticity and reading musical notation

In contrast to working memory and executive functioning, automaticity mostly bypasses the PFC as memory is instead retrieved in a non-conscious, immediate manner. Two forms of automaticity are distinguished, early-stage and procedural, and both are foundational in circumventing the PFC in accessing memory and actions. Bypassing the PFC helps to prevent cognitive overload from multitasking or task switching, thereby enabling efficient, effortless performance (Servant et al., 2018).

Automaticity can only be achieved after initial engagement with WM and executive functioning; practice and repetition of skills or information initially learned through PFC mechanisms will move toward long-term memory storage (Servant et al., 2018). Moreover, repetitive practice substantially increases the speed and accuracy of automatic content retrieval (Wilkins and Rawson, 2011). Musicians often exhibit degrees of automaticity when performing, with studies revealing reduced activity in the dorsolateral prefrontal cortex during such instances (Tan et al., 2024). This phenomenon is known as the “transient hypofrontality hypothesis” (Dietrich, 2004). Instead of the PFC being activated while performing, posterior areas of the brain become more operational, with regions such as Broca's area showing more activity (Jeon and Friederici, 2015; Liao et al., 2024a).

2.3.1 Early-stage automaticity: instance-based retrieval

Early automaticity emerges when repeated exposure allows specific notational patterns to be retrieved as instances (Logan, 1998). This type of automaticity remains partly dependent on fronto-parietal networks and is sensitive to task complexity. In notation reading, early automaticity supports rapid recognition of familiar rhythmic or scalar patterns.

2.3.2 Procedural automaticity: sensorimotor integration

With extended deliberate practice (Ericsson, 2008), automaticity transitions into proceduralized sensorimotor routines consolidated in cortico-striatal–cerebellar circuits (Ashby et al., 2010). This later-stage automaticity enables parallel processing, freeing predictive systems to operate ahead of real-time performance demands. Procedural automaticity is the form most relevant to bi-temporal music reading. With repetitive practice being the foundation of much music pedagogy (Persson, 1996; Stambaugh, 2011), this time-tested method is crucial for achieving high levels of automaticity in music performance. To understand how repetitive practice fosters automaticity, we now examine the principles of Hebbian learning.

2.4 Hebbian learning and automaticity

Hebbian learning theory posits that concurrent neural activity results in lasting synaptic strengthening (Hebb, 1949), which partially explains how motor skill development in instrumental music performance is learned to a point of automaticity. Known also by the phrase “neurons that fire together, wire together,” coined by Carla Schatz of Harvard University to describe Hebb's explanation of his theory (Collins, 2017), Hebbian learning is independent of any feedback-related synaptic strengthening, instead developing more slowly, over time. An example is procedural practice, which requires repetition to establish and then consolidate the necessary skills (Ashby et al., 2007), a common preparatory approach for musicians.

Hebbian learning, which essentially describes a mode of brain plasticity, may be contextualized in music as follows: when a musician practices reading notation and playing, the simultaneous activation of visual neurons (responding to the notes), motor neurons (controlling finger movements), and auditory neurons (responding to the sounds or imagined pitches) causes the affected neural circuits to strengthen their connections. Over time, these repeated associations forge a tight sensorimotor linkage between seeing a note and executing the corresponding action. Cognitive neuroscience frameworks describe this as a form of common coding (Prinz, 1997) between perception and action: a joint representation emerges when specific actions are repeatedly paired with specific sensory events. In music, a given note on the staff becomes inextricably associated with the finger movement and the sound that note produces. The result is that perceiving the stimulus (the written note) automatically activates the motor program to play it and the auditory representation of its sound.

Hebbian plasticity explains how reading notation becomes ingrained at the neural level. Units of neurons, or cell assemblies, form new associations to create “chunks” of information (Lörch et al., 2023). Chunking plays a foundational role in the formation of procedural routines in music learning. Research in Neurologic Music Therapy (Thaut et al., 2015) demonstrates that rhythmic and melodic grouping enhances temporal prediction, reduces working-memory load, and stabilizes motor sequencing. Incorporating this perspective strengthens the theoretical link between Hebbian learning and the emergence of multi-note “units” that musicians process as single functional entities. Each rehearsal that links a note on the page with a finger movement and, if a musician is using audiation, a relative understanding of pitch spatiality, reinforces the synaptic connections among the corresponding visual, auditory, and motor neurons. Eventually, these neural pathways become so strong that the presence of one element (e.g., a note written on a staff) may be sufficient to trigger the entire synaptic connection. When this occurs, the optimal outcome for performance is facilitated, allowing the musician to perform in a state of automaticity. Because automaticity is generally understood as a process that allows actions to be performed in parallel and without attention (Schneider and Shiffrin, 1977), fewer mistakes are made. Automaticity also enables musicians to respond to unexpected stimuli during a performance or to focus on elements of stagecraft to enhance audience engagement. Furthermore, chunking mechanisms directly support the bi-temporal model by allowing larger perceptual units to be prepared in Time Zone 1 while Time Zone 2 handles execution.

3 Neural foundations of music-notation processing: evidence and limitations

To lay a foundation for understanding the present discussion, we must first examine the current perspectives, research, and theories related to the neural basis of interpreting music notation and related understandings in the domains of neuroscience and psychology. This will be followed by a systematic review of brain imaging studies that incorporate aspects of reading musical notation. Through this review, I highlight that for the most part, the foci within the stimulus materials for musician-participants of these studies do not replicate real-world applications of reading and interpreting musical notation. A dissonance exists between the data presented in some of these studies and the assumptions made about the neural bases of reading music in real performance contexts. This dissonance, in part, offers a rationale for the theory presented in this article, as it is grounded in the common, everyday experiences of performing in a musical context.

3.1 Neural mechanisms supporting visual, predictive, and motor pathways

When engaged in music performance, cognitive processes relating to memory and motor function are fundamentally interconnected (Bigand et al., 2000; Ho et al., 2003; Zimmerman and Lahav, 2012). The various roles of motor function components are well-understood in the context of learning music (Edwards and Hodges, 2008; Wilson, 1986). Motor function learning is linked to procedural memory (Ashby et al., 2003), with automaticity considered to be an outcome of deliberate music practice (Ashby and Crossley, 2012). These include the relationships between large and small muscles, coordination when playing an instrument, proprioception—the physical manifestation of rhythm—and the execution of motor function after conceptualizing mental plans for action (Palmer and Meyer, 2000; Sidnell, 1986).

The interpretation of music notation in the context of traditional Western art music relies on the visual system1. As the eyes scan the score, visual cortices process the basic features of the symbols (lines, spaces, and note heads). In musically literate individuals, higher-order visual areas develop specialized responses to notation. Functional magnetic resonance imaging (fMRI) studies have shown that reading music engages the occipito-temporal cortex, particularly in the right hemisphere, in a manner distinct from reading text or numbers (Schön et al., 2002).

Saccades, the rapid eye movements that occur between fixations, are a crucial component of visual information processing during sight reading. In an influential study, Goolsby demonstrated that highly proficient pianists did not confine their gaze to individual notes but instead employed dynamic saccadic movements across the score, suggesting that their processing extended beyond immediate visual input (Goolsby, 1994). By contrast, less experienced readers displayed longer fixations on discrete notes, suggesting a more localized, effortful mode of visual engagement. These findings highlight significant differences in how visual stimuli are processed as a function of skill level, with advanced performers relying less on conscious decoding and more on anticipatory processing.

Subsequent research has reinforced this view. Waters, Townsend, and Underwood found that skilled pianists exhibited shorter fixation durations and a wider perceptual span compared to novices, enabling them to preview and integrate upcoming material with minimal disruption to performance (Waters et al., 1998). Similarly, Furneaux and Land observed that expert pianists made more forward-directed saccades, reflecting a predictive mechanism that supports real-time motor execution (Furneaux and Land, 2000). More recently, Penttinen et al. demonstrated that proficient sight-readers engaged in anticipatory fixations, allowing them to maintain fluency even when presented with unfamiliar rhythmic and melodic material (Penttinen et al., 2014), underscoring the role of predictive eye-movement strategies.

Recent research on parafoveal processing provides a complementary mechanism for read-ahead behavior. Pan et al. demonstrated that readers extract lexical information from parafoveal regions before saccades occur, suggesting a temporal separation between perceptual intake and motor preparation (Pan et al., 2021). Applied to music reading, parafoveal processing enables musicians to preview upcoming notes, patterns, or structural cues before initiating the eye movement. This mechanism aligns naturally with the proposed bi-temporal model, as saccades support spatial transitions across the score, detecting familiar patterns in the notation and allowing parafoveal processing to anticipate several beats ahead while real-time execution of the music performance takes place.

From a cognitive perspective, these findings of visual scanning and saccades can be situated within the framework of working memory. According to Baddeley's revised model, the phonological loop and visuospatial sketchpad support the short-term retention of symbolic information (Baddeley, 2012); however, their limited capacity necessitates the development of automated retrieval processes for fluent performance. Skilled sight-readers appear to bypass some of these capacity constraints through proceduralized, automatic processing (Ashby et al., 2010), reducing reliance on the prefrontal attentional system. The automaticity hypothesis suggests that repeated sensorimotor coupling of visual notation with motor output strengthens associative networks, allowing musicians to respond reflexively to notated stimuli while simultaneously allocating cognitive resources to higher-order planning and error monitoring (Ashby et al., 2010).

3.2 Cognitive processing models of interpreting music notation

In describing how musicians can read and translate music notation, one of the predominant theories guiding music education is Edwin Gordon's audiation framework, a component of Gordon's overall Music Learning Theory (Gordon, 2011). In an earlier outline of his theory, Gordon expressed that “audiation is to music what thought is to language” (Gordon, 1999, p. 42), further clarifying that aural perception is distinct from audiation. Moreover, the act of decoding notation is different from engaging in audiation, as the former does not prescribe an inner sense of hearing.

Psychological research has suggested that highly trained musicians process written music holistically and automatically (Wong and Gauthier, 2010a). Holistic processing on an automatic level is typically associated with facial recognition (Farah et al., 1998) and visual elements unique to a particular area of expertise (Curby and Moerel, 2019). Holistic processing, defined as the failure of selective attention (Curby and Moerel, 2019; Wong and Gauthier, 2010a), shares some aspects with the principles of Gestalt perception (Wertheimer, 1912); however, there are some differences between the two processing models. Holistic processing refers to perceiving complex stimuli as integrated wholes, rather than as individual elements. Gestalt principle, on the other hand, describes how visual components may be grouped intuitively to form the impression of a whole. In the context of music, Gestalt is considered by theorists and scholars from the perspective of time (Tenney and Polansky, 1980) and the processing of structures in the musician's mind (Reybrouck, 1996; Terhardt et al., 1987).

When examined in the context of reading notation, Wong and Gauthier found that musicians with higher levels of expertise could distinguish relevant information using brain regions distinct from those of novice music readers (Wong and Gauthier, 2010a). For novice music readers, neural responses were observed in the right fusiform face area (rFFA), the brain region to which holistic processing, such as facial recognition, is attributed (Farah et al., 1998; Maurer et al., 2002; Wong and Gauthier, 2010a). However, for expert music readers, activity in the rFFA was negatively correlated with years of experience, indicating that greater expertise in music reading was associated with holistic processing and automaticity occurring in brain regions other than rFFA (Wong and Gauthier, 2010a). In this study, examples of holistic processing among novice music readers were “strategic” in novice musicians, requiring attention to individual notes for understanding. In contrast, experienced musicians displayed evidence of abilities to identify constituent parts of music notation as well as patterns and sequences similar to that of facial recognition.

Schön et al. found a specific activation focus at the right occipito-temporal junction when pianists read musical notes, a region they likened to a musical analog of the visual word form area for text reading (Schön et al., 2002). This suggests that, through training, the brain develops a specialized area for recognizing visual symbols associated with musical notes. Additionally, musical literacy can induce plastic changes in visual regions, with short-term training in reading music enhancing fusiform gyrus responses to notation (Stewart, 2005a). Such findings indicate that the visual cortex not only detects the shapes of notes but also becomes tuned to the unique patterns of musical symbols in experienced readers.

3.3 Systematic review of neuroimaging studies with a focus on music notation

A problem currently present in the literature regarding the neuronal bases of reading music notation is that, apart from a few examples, most brain imaging studies investigating this topic focus on single parameters of music notation reading. This contradicts real-world music performance, which involves many specific components of stimulus-response processing (Gruhn and Rauscher, 2007; Zatorre, 2012). Another issue is the limited research available on prospective planning, an element of performance critical to successful outcomes when reading music from notation. A third concern is the lack of understanding of how musicians can simultaneously operate the extensive neural networks necessary for music performance (Altenmüller and Furuya, 2015).

Imaging studies have yielded significant data on how music affects biological structures of the brain. Such studies encompass a wide range of elements related to music, including listening to music, playing an instrument, interpreting music notation, near and far transfer of skills, and long-term musical training. A large body of research exists on how learning and performing music influence neurobiological structures and promote synaptic plasticity (Proverbio and Sanoubari, 2024), as well as how reading small examples of music notation may be represented hemispherically in the brain (D'Anselmo et al., 2015). These data are essential in their contributions to understanding the neural bases for reading notation. However, there is a scarcity of information regarding how cognitive processes involved in interpreting musical notation operate in the real-world context of a musician simultaneously negotiating numerous actions while prospectively planning for future task operations.

In recent literature, brain imaging studies that incorporate the interpretation of music notation primarily focus on individual parameters, such as pitch or rhythm. While there are instances where pitch and rhythm are combined in music notation stimuli, there may be constraints on complexity, such as using simplified rhythms like quarter and eighth notes. However, the real-world application of reading music for a musician encompasses a vast array of activities that necessarily integrate an expansive neuronal network, while also requiring a high level of attentional focus in the PFC region of the brain. Given the complex integration of tasks necessary to perform music, the neural processes of interpreting music notation need to be evaluated in the context of simultaneous task action alongside multiple sensory stimuli. This sentiment is echoed by Schön and Besson, who argue that “since melody and harmony both contribute to the rhythmic organization of a musical work, and since neither melody nor harmony can be activated without rhythm, the three must be regarded as inseparably linked” (Schön and Besson, 2002, p. 868) and posit that for music to be adequately represented in brain imaging studies, these elements must be integrated.

4 Methods and materials

To investigate cognitive processes as they relate to neural bases in interpreting music notation, a review of brain imaging studies of music notation was conducted. This review followed the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA 2020) guidelines to ensure transparency and reproducibility in the identification, screening, and inclusion of studies examining the neural correlates of reading music notation (see Figure 1). Employing a systematic approach, defined as “a review of existing research using explicit, accountable, rigorous research methods” (Gough et al., 2017, p. 4), I searched the following electronic databases: Google Scholar, Scopus, Web of Science, and EBSCO, using specific search terms to narrow the results to a high level of relevance. To determine the relevance to the present investigation, information was extracted from the abstracts, with unrelated studies discarded.

Figure 1
Flowchart illustrating the identification of studies via databases. Records identified from four databases totaled 278, with 183 removed as ineligible. Ninety-five records were screened, with 56 duplicates removed. Thirty-nine reports were sought, all were retrieved. Thirty-nine reports were assessed, with two excluded. Thirty-seven studies were included in the review.

Figure 1. PRISMA 2020 flow diagram. This work is licensed under CC BY 4.0. To view a copy of this license, visit https://creativecommons.org/licenses/by/4.0/. *Reports marked ineligible by the author, not automation tools.

The inclusion criteria were as follows: (1) that cognitive functions relating to interpreting music notation were involved in the study; (2) that the tests performed in the studies incorporated responding to the visual stimulus of music notation; (3) that the studies were published in peer-reviewed journals or books (dissertations and theses were not considered); (4) that the studies were of original content, and not, for example, literature reviews of previous studies (i.e., primary studies); and (5) that the results returned were either open access or able to be viewed via institutional subscriptions. The search queries were: (1) fMRI music reading; (2) “fMRI” and “music reading;” (3) “brain imaging” and “music notation;” and (4) “brain imaging” and “music reading.” Analysis was limited to the first 30 results for each search term in each database, resulting in a possible return of 278 results to investigate. After applying the inclusion criteria to each of the 278 articles, 95 listings (34.17%) were identified as meeting the requirements for this review (see Table 1).

Table 1
www.frontiersin.org

Table 1. Literature search results: brain imaging and reading music notation.

An investigation was conducted on publications deemed relevant to this study that summarized the following elements: the brain imaging technique used, the musical parameters under investigation (e.g., pitch, rhythm), and the complexity of these parameters. These analyses served to ground the hypothesis central to this article by providing a perspective on conditions and contexts for musicians who participate in brain imaging studies that incorporate reading music notation.

A narrative synthesis approach was chosen because the studies included in this review demonstrated heterogeneity across imaging modality, task design, and participant expertise. These variations presented limitations for calculating meaningful comparisons, rendering a statistical meta-analysis inappropriate. The narrative synthesis approach preserved the contextual, conceptual, and methodological nuances of each study. This approach facilitated the identification of convergent patterns in neural activation, the evaluation of discrepancies arising from divergent task structures, and the critical examination of ecological validity in existing research on music-notation processing. Furthermore, given that this review aims to develop and refine a theoretical model of bi-temporal cognitive processing, narrative synthesis offered the necessary flexibility to connect empirical results with broader cognitive and neuroscientific frameworks.

5 Results

Of the 95 relevant articles identified across the databases, numerous instances of duplication were found. Among these articles, 37 were unique (38.94%), with 22 represented in two or more different databases. The remaining 15 articles appeared in only one database across the 16 possible search iterations (see Table 2).

Table 2
www.frontiersin.org

Table 2. Article summary.

In brain imaging studies on cognitive processing during music notation reading, simple representations of pitch (notes) and duration (rhythm) were commonly used as stimulus materials. This review of the literature focused on pitch and rhythm, either as individual items for study using brain imaging techniques or in simple combinations. Table 3 presents the primary focus and complexity of music notation across the 37 studies, along with the brain imaging techniques employed.

Table 3
www.frontiersin.org

Table 3. Review of brain imaging studies involving music notation reading.

While some of these studies engaged with musical stimuli representing performance contexts as experienced by professional musicians, the simplicity of most of the musical notation calls into question what we really know about brain mapping and cognitive processing when reading music. This simplicity reduces the need for conscious effort for participants with high levels of literacy, and importantly, negates the opportunity to discover neural correlates of simultaneous task function when reading music.

In terms of content explored in the collection of literature, there were 18 articles focused on pitch (48.64%), 14 articles focused on both pitch and rhythm parameters (37.83%), three articles on rhythm alone (8.10%), and two articles that did not specify a specific focal point. Figure 2 represents the breakdown of article focus across the sample.

Figure 2
Pie chart showing four categories. Pitch 48.64% (light blue), pitch and rhythm is 37.83% (dark blue), rhythm is 8.10% (light green) and unspecified is 5.43% in dark green).

Figure 2. Focus of notational elements across the sample literature.

To summarize the representation of complexity or simplicity in the literature sample, Figure 3 shows that out of the 37 articles examined, 26 (70.27%) used simple music notation as the visual stimuli. In the simple to moderate category, three articles used slightly more complex or layered content to simulate notation stimuli (8.10%). There were six articles that used moderate-to-difficult notational stimuli (16.21%), and two articles referenced piano pieces from the broader repertoire (5.40%).

Figure 3
Pie chart showing the distribution of musical complexity in brain-imaging studies of reading music notation. Simple notation is 70.27% (dark blue), moderate to difficult is 16.21% (green), slightly more complex is 8.10% (purple), and unspecified is 5.40% (light blue).

Figure 3. Levels of complexity of music notation used in sample literature.

6 Discussion

The convergence of literature from the domains of psychology, neuroscience, and music education highlights the important role of automaticity in instrumental performance and the neural mechanisms underpinning music reading. Overall, the existing literature provides partial but fragmented support for a bi-temporal account of music-notation processing. While behavioral and eye-movement studies are broadly consistent with predictive, dual-pathway processing, neuroimaging evidence remains constrained by limited task complexity and insufficient attention to prospective planning. These limitations do not undermine the model but instead motivate its theoretical necessity, which is more broadly addressed in Section 6.4.

One area that appears to be well-supported in the literature is that, far from being a simple consequence of repetition, automaticity emerges through the interaction of domain-specific cognitive processes and Hebbian-driven plasticity in cortical and subcortical circuits (Jäncke, 2012). Neuroimaging studies demonstrate that reading music notation recruits a distributed network including the visual cortex, superior temporal gyrus, premotor cortex, and supplementary motor areas, which together facilitate the transformation of visual symbols into motor commands and auditory imagery (Stewart et al., 2003b; Schön and Besson, 2002). With sustained practice, musicians transition from a conscious, declarative decoding of notation to the automatic engagement of proceduralized routines, a process associated with strengthened connectivity between the dorsolateral prefrontal cortex, basal ganglia, and cerebellum (Bangert et al., 2006; Zatorre et al., 2007). Taken together, the evidence indicates that music-notation reading recruits distributed visual, auditory, and motor networks that partially overlap with those involved in text reading but diverge in their reliance on predictive audiation and sensorimotor integration, reflecting the performative demands unique to musical notation.

6.1 Understanding the limitations of brain imaging studies

There are complexities associated with accurately interpreting brain scans of the human brain. The stochastic nature of activity in biological processes can produce extraneous data that can be analyzed but not necessarily predicted, and may arise due to intrinsic factors, such as normal processes of protein transfer across cell membranes, or extrinsic factors like increased stress hormones resulting from environmental changes (Eling et al., 2019). When measuring dynamic brain activity, such extraneous biological activity, or “noise,” can interfere with the fidelity of the data, potentially resulting in a misrepresentation of neuronal activity. It is therefore crucial that baseline measurements are taken as part of all studies that isolate specific variables for analysis, as well as the resting state of the participants.

Adjacent to the problem of biological noise in brain imaging is the possibility of elevated correlations resulting from generalizations in contemporary brain research analysis. Bandettini points out that a current challenge in referring to statistical mapping structures to interpret analytical findings is that “the maps are not appropriately normalized to multiple comparisons” (Bandettini, 2020, p. 184), which can lead to incorrect analyses when statistical maps are used to form a baseline comparison. Additionally, elevated correlations may be derived when studies are conducted that do not account for independent variables. An example of this is described by Vul et al., in their review of 55 brain imaging studies (53 studies, after two were deemed not to meet the criteria), where they identified non-specific explanations for the conditions under which voxel activations were selected for data analysis (Vul et al., 2009). Specifically, the authors noted that misrepresentations occurred when authors of the 53 studies reported high levels of correlation between activations in specific brain areas and stimuli that provoked responses in emotional, personality, and social cognition categories.

Considering the risks of biological noise interference or unintentional misinterpretation of brain imaging data, it is essential to isolate specific learning tasks or actions when conducting studies of the human mind. Also, because statistical maps informing the interpretation of data may not always be reliable (Bandettini, 2020), responsible practice in obtaining replicable and authentic imaging information must be at the forefront of studies. From there, gradually increasing complexities in music notation stimuli could be possible and may yield results previously undiscovered.

6.1.1 Determining levels of expertise for musician-participants

Another constraint in brain imaging studies is bias, specifically in relation to the level of competency of musicians under study, and how this competency is interpreted. Many studies in this body of literature refer to subjects within their protocols as “trained musicians,” substantiating this by outlining the number of years the participants have taken lessons on their instrument (Brown and Penhune, 2018; Giovannelli et al., 2020; Kawasaki and Hayashi, 2022; Liao et al., 2024b; Lu et al., 2022, 2019; Nichols and Grahn, 2016; Schön et al., 2002; Schön and Besson, 2002; Simoens and Tervaniemi, 2013; Wakita, 2016; Wong and Gauthier, 2010a), or if they perform professionally or study at a conservatoire (Bouhali et al., 2017, 2020; Brodsky et al., 2008; de Manzano and Ullén, 2012; Endestad et al., 2020; Hoppe et al., 2014; Lee and Lei, 2012; Liao et al., 2024a,b; Meister et al., 2004; Mongelli et al., 2017; Paraskevopoulos et al., 2014b; Proverbio and Valtolina, 2025; Ross et al., 2013; Simoens and Tervaniemi, 2013; Stewart, 2005a,b). If a study focuses on the entrainment aspects of novel musical information for non-musicians (Stewart et al., 2003a), a restriction on the time spent previously learning an instrument may be stated (Chang et al., 2025; Karagiorgis et al., 2021; Stewart et al., 2003b) or a declaration may be made that the participants had not received formal training prior to the study (Lee and Wang, 2011; Paraskevopoulos et al., 2014a).

However, even at the professional level, there is a variance in skill competencies, which is why musicians choose to specialize in certain areas. Furthermore, conservatorium-trained musicians who become career musicians in orchestras may refine their focus to either traditional classical music, playing in a philharmonic or symphony orchestra. If they specialize in contemporary classical music, they are likely to play in smaller, chamber ensembles, often with more challenging and soloistic repertoire. The distinction between even just these two styles of music—two styles out of several hundred—raises a critical question about musical language and what a musician may be familiar with in terms of patterns and paradigms. Some studies investigating how musicians interpret music notation claim to use lesser-known etudes or works composed by specialists as an unknown stimulus for musician participants, yet the underlying language within such music is steeped in predictability. When adhering to a style associated with traditional forms of classical music, detectable patterns and compositional characteristics can be distilled into easily identifiable scale and arpeggio-based fragments. Altenmüller and Furuya describe this phenomenon as musicians having a “similar acculturation due to the canonical nature of their training” (Altenmüller and Furuya, 2015, p. 4). Therefore, using music notation that displays stylistic characteristics typical of the classical genre does not provide a novel stimulus of music. This crucial point potentially circumvents the aims of studies that investigate the cognitive processing of unrehearsed musical material as a central focus. Instead, data derived from a study that relies on classical-archetypal notational material will likely measure the identification of deeply familiar patterns, forged from a strong procedural memory that has been consolidated over years of practice, rather than responses to “novel” musical stimuli.

6.1.2 Different levels of musical ability and implications for data reliability

The studies in this literature review offered varying levels of detail regarding the musician-participants' levels of experience and competencies. While some studies included detailed supplementary material with fulsome information about the participants validating their expertise (Proverbio and Valtolina, 2025), others considered musical training in excess of one year to be sufficient for a participant to be categorized as a musician (Nichols and Grahn, 2016). In a study using EEG, hEOG, and vEOG, Proverbio and Valtolina acknowledged the limitations of available musical stimuli for advanced musicians (Proverbio and Valtolina, 2025). Discounting the use of rhythmic tests, such as the Montreal Battery of Evaluation of Musical Abilities, in baseline testing because they were too simple for the experienced musician participants, the authors opted to incorporate more complex notational stimuli in the tests, aligning more closely with real-world applications of performing from notation for highly trained musicians.

Stewart et al. presented a robust publication in which musically naïve participants were under study for evidence of skill development (Stewart et al., 2003b). To quantify the level of performance target, participants were given music notation and theory stimuli at a Grade 1 level (UK: Associated Board of the Royal Schools of Music). Because participant progress was not uniform, some extra tuition was required to achieve the baseline level of proficiency needed to complete the testing. Another example of considered attention to the variable of musical competency is seen in Wong and Gauthier's study. Here, the authors acknowledge that the automatic holistic processing of notation observed in musician participants was potentially representative of “an increased tendency to process relative positions of notes in music sequences, as one becomes more proficient with musical notation” (Wong and Gauthier, 2010a, p. 549), noting the correlation of years of musical training and fluency in reading notation. Furthermore, Wong and Gauthier accurately pointed out in this study that levels of fluency in reading notation are variable, even within populations of expert musicians.

6.2 The case for real-world replication in brain imaging studies

There are many constituent parts to music notation that extend beyond these two elements of pitch and duration, which were highly represented in Table 3. A piece of music will commonly include elements such as tonality (the key to which the piece belongs), dynamics (volume), and articulation (different ways in which a note may be accented, shortened, or stressed). Furthermore, there may be significant modulations throughout a piece, such as tempo changes (speeding up or slowing down as indicated by the instructions included on the sheet music), key signature changes (alterations to not only how many sharps or flats the musician needs to remember but also which scale or tonality they are working in), meter (the time signature that instructs the musician on how to count the beats), the structure (repeat signs and other instructions that inform the musician whether they are to repeat certain sections), and clef changes (the symbol at the beginning of a piece of music that informs the musician how to interpret the pitch placement on the five-lined staff). Table 4 provides a non-exhaustive overview of the parameters a musician simultaneously manages when performing.

Table 4
www.frontiersin.org

Table 4. Parameters of music notation: examples.

Clef changes are particularly interesting because a musician who is switching between, say, a treble clef and an alto clef, will need to reconfigure their mental positioning of the notes they are looking at. In such an example, what visually presents as a B4 in treble clef becomes a seventh lower in the alto clef (C4). Figure 4 demonstrates the effect of a clef change from treble to alto: the same note is represented in two vastly different ways.

Figure 4
Music notation displaying two staves. The top staff has a treble clef with a middle C note. The bottom staff has an alto clef with a middle C note.

Figure 4. Example of Different Clefs in Music. The treble clef (top) instructs the musician how to interpret the note placement. The note to the right of the treble clef is C4. The alto clef (bottom) has repositioned the range of notes that the musician now interprets when reading the sheet music. The note to the right of the alto clef is also a C4.

When reading notation, the musician retains contextual information that determines how to play a piece (i.e., which clef is in use or which key signature to apply) through executive functioning. When a piece changes clef, key, dynamics, articulation, or other elements, the musician adapts to the new paradigm. Such shifts can only be achieved with success through prospective planning and an accompanying degree of automaticity that can sufficiently process the changing visual stimuli.

6.3 Comparing the interpretation of music notation and the written word

Reading music to a high level involves not only a deep knowledge of the visual representations of notation but also a near-immediate processing of the musical material. Interpreting music from notation is a complex task that is thought to be learned through explicit teaching (Hébert and Cuddy, 2006). A prevailing difference between the interpretation of musical notation and written text is that notation requires the decoding of multiple events and instructions that co-occur, whereas reading words on a page is operationalized in a linear, sequential manner (Sloboda, 2004). Furthermore, reading music notation requires a musician to perform what they see, assessing the aural output in real-time while simultaneously reading notation that has yet to be played. According to Wöllner and Williamon, musicians can construct “sonic images,” which allows for the anticipation of aural stimuli, whether or not auditory feedback is available (Wöllner and Williamon, 2007). Musicians with high levels of expertise conceptualize these sonic images, and when in the act of reading notation, this process is commonly referred to as audiation, that is, the “ability to hear and give meaning to music when sound is not physically present or may never have been physically present” (Gordon, 2011, p. 10). Audiation is a cornerstone of accurate musical performance, and its use contributes to the prospective planning that is necessary when playing an instrument or singing.

6.4 Introducing the theory of bi-temporal cognitive processing in reading music notation

Several factors operationalize the neural basis of reading music notation. First, the capacity for musicians to enact prospective and retrospective planning simultaneously (Tsao et al., 2022) enables them to perform in real time while reading ahead in the score. This act engages working memory and predictive coding mechanisms in the prefrontal and parietal cortices (Waters et al., 1998; Palmer and van de Sande, 1993). Second, purposeful, repetitive practice over a long period will lead to Hebbian learning as the musician develops their skill set. Maintaining this practice will eventually result in automaticity. Thus, Hebbian learning and the emergence of procedural automaticity provide the neural conditions under which present-focused execution can proceed in parallel with future-oriented prediction, resolving the apparent multitasking paradox in skilled music reading. Third, the act of reading music requires the musician to read ahead and audiate the notation while concurrently performing notation read only moments earlier. This means that a bi-temporal cognitive processing paradigm is central to the success of musical performances: the musician operates in two time zones, producing a single musical outcome. The proposed time zones as seen in Figure 5 reflect functional, temporal orientation rather than discrete neural substrates.

Figure 5
Diagram illustrating brain network interactions across two time zones: real-time and anticipatory. It shows the ventral and dorsal attention networks interacting with the auditory and visual cortex, linked by the executive control network (ECN) to the prefrontal, parietal, and premotor & motor cortex. These are connected to the basal ganglia and cerebellum, leading to outcomes of automaticity and Hebbian learning. Pathways are indicated by colored arrows representing different processes.

Figure 5. Bi-Temporal Cognitive Processing Theory for Music Notation Reading. Automaticity in music performance reflects the integration of cortical and subcortical systems operating in dual temporal domains (Zatorre et al., 2007). Premotor-motor circuits mediate real-time execution (Scott, 2004; Shenoy et al., 2013; Porter and Lemon, 1995), while anticipatory planning engages prefrontal and parietal regions (Miller and Cohen, 2001); these processes converge in the basal ganglia and cerebellum (Ashby et al., 2010), where Hebbian plasticity consolidates proceduralization and error correction, enabling fluent, predictive performance (Wolpert et al., 1998). This graphic representation is the author's own work.

Figure 5 illustrates how automaticity in music performance emerges from distributed cortical and subcortical interactions that operate simultaneously in two cognitive time zones. The visual cortex initiates processing by decoding notation, which is then projected into distinct but overlapping functional streams. In Time Zone 1, information is initially processed through the ventral attention network before moving through the premotor and motor cortices, supporting real-time action response in performance. In Time Zone 2, the dorsal attention network is activated along with the prefrontal and parietal cortices, which engage working memory, predictive coding, and spatial anticipation. This network distribution enables musicians to plan prospective actions while monitoring recent performance events. These cortical processes converge on subcortical structures: the basal ganglia, which support habit formation and proceduralization, and the cerebellum, which refines timing and error correction. Through repetitive, purposeful practice, Hebbian plasticity strengthens these networks, consolidating the mapping of visual symbols onto motor and auditory representations. The outcome is automaticity in reading music notation, characterized by the ability to perform fluently in real time while anticipating future musical events. This bi-temporal framework provides a neural basis for the dual processing demands unique to skilled music performance and the interpretation of notation.

7 Conclusion

This theoretical article follows a line of inquiry into the underpinnings of interpreting musical notation within the context of authentic application, the differences between reading music and the written word, the concept of bi-temporal focus and engagement, and Hebbian learning as a basis for understanding automaticity in music performance. A systematic review of the literature examining standard practices and contexts for brain imaging studies provided an essential perspective for discussion. Here, the dissonances between real-world music performance contexts and the musical stimuli found in brain imaging studies were outlined. Connections were drawn between the distinct elements of reading music, aiming to consolidate an understanding of how the musical mind interprets notation.

The theory central to this article is a considered evaluation of cognitive processes that frame the neural basis of interpreting notation in the context of musical performance—a cognitive framework for reading music that hypothesizes the simultaneous dual processing of notational stimuli across two time zones. The pedagogical benefits of deliberate practice as a facilitator of transitioning from controlled to automatic notation processing are significant; understanding this phenomenon in depth might inform future teaching practices. In the fields of psychology and neuroscience, this theory has the potential to guide future research into the neural correlates of reading musical notation, as well as other studies investigating temporally driven cognition.

7.1 Testing the theory: recommendations for future study

To interrogate the plausibility of the theory, brain imaging studies that combine the many facets of music performance along with challenging repertoire as the visual stimuli might reveal new perspectives on reading notation. A framework for a study could include secondary tasks that involve temporal activities, measuring neural structures for time, prospective, and retrospective planning in contexts other than music. Comparing these data with scans of musicians in the act of performance might provide evidence supporting the argument that the cognitive processing of music occurs in a bi-temporal fashion. To test the bi-temporal model directly, future studies must be capable of isolating simultaneous execution and prediction rather than treating music reading as a unidimensional decoding task.

Another aspect of this theory to explore is that of Hebbian learning and automaticity. Isolating these factors in a longitudinal study before conducting an intervention study involving bi-temporal cognitive processing activities may be important in ascertaining the skill level required to test the theory authentically. Obtaining these data could define what constitutes a “trained” musician, directing evaluations away from the number of years spent learning an instrument or a working knowledge of notes and musical performance, toward fundamental skills of automaticity resulting from deliberate and repetitive practice.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

KH: Writing – original draft, Writing – review & editing.

Funding

The author(s) declared that financial support was not received for this work and/or its publication.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

1. ^Music notation is not limited only to a visual Western traditional classical music model. For example, sheet music is available for blind people through Braille systems, which provides comprehensive information for the musician performing the music. However, for the purposes of this article, the focus will be on the visual interpretation of notation that is widely used through the media of either printed sheet music or PDF document viewed on a computer or tablet.

References

Altenmüller, E., and Furuya, S. (2015). “Planning and performance,” in The Oxford Handbook of Music Psychology, ed S. Hallam (Oxford: Oxford University Press). doi: 10.1093/oxfordhb/9780198722946.013.32

Crossref Full Text | Google Scholar

Ashby, F. G., and Crossley, M. J. (2012). Automaticity and multiple memory systems. Wiley Interdiscip. Rev Cogn. Sci. 3, 363–376. doi: 10.1002/wcs.1172

PubMed Abstract | Crossref Full Text | Google Scholar

Ashby, F. G., Ell, S. W., and Waldron, E. M. (2003). Procedural learning in perceptual categorization. Mem. Cognit. 31, 1114–1125. doi: 10.3758/BF03196132

PubMed Abstract | Crossref Full Text | Google Scholar

Ashby, F. G., Ennis, J. M., and Spiering, B. J. (2007). A neurobiological theory of automaticity in perceptual categorization. Psychol. Rev. 114, 632–656. doi: 10.1037/0033-295X.114.3.632

PubMed Abstract | Crossref Full Text | Google Scholar

Ashby, F. G., Turner, B. O., and Horvitz, J. C. (2010). Cortical and basal ganglia contributions to habit learning and automaticity. Trends Cogn. Sci. 14, 208–215. doi: 10.1016/j.tics.2010.02.001

PubMed Abstract | Crossref Full Text | Google Scholar

Baddeley, A. D. (2012). Working memory: theories, models, and controversies. Annu. Rev. Psychol. 63, 1–29. doi: 10.1146/annurev-psych-120710-100422

PubMed Abstract | Crossref Full Text | Google Scholar

Baddeley, A. D., and Hitch, G. J. (1974). “Working memory,” in Psychology of Learning and Motivation, ed G. H. Bower, Vol. 8 (New York, NY: Academic Press), 47–89. doi: 10.1016/S0079-7421(08)60452-1

PubMed Abstract | Crossref Full Text | Google Scholar

Bandettini, P. A. (2020). fMRI. Cambridge, MA: The MIT Press. doi: 10.7551/mitpress/10584.001.0001

Crossref Full Text | Google Scholar

Bangert, M., Peschel, T., Schlaug, G., Rotte, M., Drescher, D., Hinrichs, H., et al. (2006). Shared networks for auditory and motor processing in professional pianists: evidence from fMRI conjunction. NeuroImage 30, 917–926. doi: 10.1016/j.neuroimage.2005.10.044

PubMed Abstract | Crossref Full Text | Google Scholar

Bigand, E., McAdams, S., and Forêt, S. (2000). Divided attention in music. Int. J. Psychol. 35, 270–278. doi: 10.1080/002075900750047987

Crossref Full Text | Google Scholar

Bouhali, F., Mongelli, V., and Cohen, L. (2017). Musical literacy shifts asymmetries in the ventral visual cortex. NeuroImage 156, 445–455. doi: 10.1016/j.neuroimage.2017.04.027

PubMed Abstract | Crossref Full Text | Google Scholar

Bouhali, F., Mongelli, V., and de Schotten, M. T,. Cohen, L. (2020). Reading music and words: the anatomical connectivity of musicians' visual cortex. NeuroImage 212:116666. doi: 10.1016/j.neuroimage.2020.116666

PubMed Abstract | Crossref Full Text | Google Scholar

Brancucci, A., Babiloni, C., Maria Rossini, P., and Luca Romani, G. (2005). Right hemisphere specialization for intensity discrimination of musical and speech sounds. Neuropsychologica 43, 1916–1923. doi: 10.1016/j.neuropsychologia.2005.03.005

PubMed Abstract | Crossref Full Text | Google Scholar

Brancucci, A., and San Martini, P. (1999). Laterality in the perception of temporal cues of musical timbre. Neuropsychologica 37, 1445–1451. doi: 10.1016/S0028-3932(99)00065-2

PubMed Abstract | Crossref Full Text | Google Scholar

Brancucci, A., and San Martini, P. (2003). Hemispheric asymmetries in the perception of rapid (timbral) and slow (nontimbral) amplitude fluctuations of complex tones. Neuropsychology 17, 451–457. doi: 10.1037/0894-4105.17.3.451

PubMed Abstract | Crossref Full Text | Google Scholar

Brodsky, W., Kessler, Y., Rubinstein, B-. S., Ginsborg, J., and Henik, A. (2008). The mental representation of music notation: notational audiation. J. Exp. Psychol. Hum. Percept. Perform. 34, 427–445. doi: 10.1037/0096-1523.34.2.427

PubMed Abstract | Crossref Full Text | Google Scholar

Brown, R. M., and Penhune, V. B. (2018). Efficacy of auditory versus motor learning for skilled and novice performers. J. Cogn. Neurosci. 30, 1657–1682. doi: 10.1162/jocn_a_01309

PubMed Abstract | Crossref Full Text | Google Scholar

Chang, Y.-H. F., Ullén, F., and de Manzano, Ö. (2025). Common brain representations of action and perception investigated with cross-modal classification of newly learned melodies. Sci. Rep. 15:16492. doi: 10.1038/s41598-025-00208-x

PubMed Abstract | Crossref Full Text | Google Scholar

Chein, J. M., and Schneider, W. (2005). Neuroimaging studies of practice-related change: fMRI and meta-analytic evidence of a domain-general control network for learning. Cogn. Brain Res. 25, 607–623. doi: 10.1016/j.cogbrainres.2005.08.013

PubMed Abstract | Crossref Full Text | Google Scholar

Collins, N. (2017). Pathways from the Eye to the Brain. Stanford Medicine Magazine, August 21, 2017. Available online at: https://stanmed.stanford.edu/carla-shatz-vision-brain/ (Accessed August 17, 2025).

Google Scholar

Curby, K. M., and Moerel, D. (2019). Behind the face of holistic perception: holistic processing of gestalt stimuli and faces recruit overlapping perceptual mechanisms. Atten. Percept. Psychophys. 81, 2873–2880. doi: 10.3758/s13414-019-01749-w

PubMed Abstract | Crossref Full Text | Google Scholar

D'Anselmo, A., Giuliani, F., Marzoli, D., Tommasi, L., and Brancucci, A. (2015). Perceptual and motor laterality effects in pianists during music sight-reading. Neuropsychologia 71, 119–127. doi: 10.1016/j.neuropsychologia.2015.03.026

PubMed Abstract | Crossref Full Text | Google Scholar

de Manzano, O., and Ullén, F. (2012). Goal-independent mechanisms for free response generation: creative and pseudo-random performance share neural substrates. NeuroImage 59, 772–780. doi: 10.1016/j.neuroimage.2011.07.016

PubMed Abstract | Crossref Full Text | Google Scholar

Dietrich, A. (2004). Neurocognitive mechanisms underlying the experience of flow. Conscious. Cogn. 13, 746–761. doi: 10.1016/j.concog.2004.07.002

PubMed Abstract | Crossref Full Text | Google Scholar

Edwards, R. D., and Hodges, D. A. (2008). “Neuromusical research: an overview of the literature,” in Neurosciences in Music Pedagogy, eds W. Gruhn, and F. H. Rauscher (New York, NY: Nova Science Publishers).

Google Scholar

Eling, N., Morgan, M. D., and Marioni, J. C. (2019). Challenges in measuring and understanding biological noise. Nat. Rev. Genet. 20, 536–548. doi: 10.1038/s41576-019-0130-6

PubMed Abstract | Crossref Full Text | Google Scholar

Endestad, T., Inge Godøy, R., Sneve, M. H., Hagen, T., Bochynska, A., Laeng, B., et al. (2020). Mental effort when playing, listening, and imagining music in one pianist's eyes and brain. Front. Hum. Neurosci. 14:576888. doi: 10.3389/fnhum.2020.576888

PubMed Abstract | Crossref Full Text | Google Scholar

Ericsson, K. A. (2008). Deliberate practice and acquisition of expert performance: a general overview. Acad. Emerg. Med. 15, 988–994. doi: 10.1111/j.1553-2712.2008.00227.x

PubMed Abstract | Crossref Full Text | Google Scholar

Farah, M. J., Wilson, K. D., Drain, M., and Tanaka, J. N. (1998). What is ‘Special' about face perception? Psychol. Rev. 105, 482–498. doi: 10.1037//0033-295X.105.3.482

PubMed Abstract | Crossref Full Text | Google Scholar

Furneaux, S., and Land, M. F. (2000). The effects of skill on the eye-hand span during musical sight-reading. Proc. R. Soc. B 266, 2435–2440. doi: 10.1098/rspb.1999.0943

PubMed Abstract | Crossref Full Text | Google Scholar

Gaser, C., and Gottfried Schlaug, G. (2003). Brain structures differ between musicians and non-musicians. J. Neurosci. 23, 9240–9245. doi: 10.1523/JNEUROSCI.23-27-09240.2003

PubMed Abstract | Crossref Full Text | Google Scholar

Gazzaley, A., and Rosen, L. D. (2016). The Distracted Mind: Ancient Brains in a High-Tech World. Cambridge, MA: MIT Press.

Google Scholar

George, E. M., and Coch, D. (2011). Music training and working memory: an ERP study. Neuropsychologia 49, 1083–1094. doi: 10.1016/j.neuropsychologia.2011.02.001

PubMed Abstract | Crossref Full Text | Google Scholar

Giovannelli, F., Rossi, S., Borgheresi, A., Gavazzi, G., Zaccara, G., Viggiano, M. P., et al. (2020). Effects of music reading on motor cortex excitability in pianists: a transcranial magnetic stimulation study. Neuroscience 437, 45–53. doi: 10.1016/j.neuroscience.2020.04.022

PubMed Abstract | Crossref Full Text | Google Scholar

Goolsby, T. W. (1994). Eye movement in music reading: effects of reading ability, notational complexity., and encounters. Music Percept. 12, 77–96. doi: 10.2307/40285756

Crossref Full Text | Google Scholar

Gordon, E. E. (1999). All about audiation and music aptitudes. Music Educ. J. 86, 41–44. doi: 10.2307/3399589

Crossref Full Text | Google Scholar

Gordon, E. E. (2011). Roots of Music Learning Theory and Audiation. Chicago, IL: GIA Publications.

Google Scholar

Gough, D., Oliver, S., and Thomas, J. (2017). “Introducing systematic reviews,” in An Introduction to Systematic Reviews, 2nd edn: eds D. Gough, S. Oliver, and J. Thomas (London: Sage), 1–18. doi: 10.53841/bpsptr.2017.23.2.95

Crossref Full Text | Google Scholar

Gruhn, W., and Rauscher, F. H. (2007). Neurosciences in Music Pedagogy. New York, NY: Nova Biomedical Books.

Google Scholar

Gunter, T. C., Schmidt, B.-H., and Besson, M. (2003). Let's face the music: a behavioral and electrophysiological exploration of score reading. Pscyhophysiology 40, 742–751. doi: 10.1111/1469-8986.00074

PubMed Abstract | Crossref Full Text | Google Scholar

Hari, J. (2022). Stolen Focus. London: Bloomsbury Publishing.

Google Scholar

Heath, K. L. (2025). Multiple Memory Systems in Instrumental Music Learning. Doctoral diss., Boston University. Boston University. doi: 10.13140/RG.2.2.21345.72804

Crossref Full Text | Google Scholar

Hebb, D. O. (1949). The Organization of Behavior. New York, NY: John Wiley and Sons.

Google Scholar

Hébert, S., and Cuddy, L. L. (2006). Music-reading deficiencies and the brain. Adv. Cogn. Psychol. 2, 199–206. doi: 10.2478/v10053-008-0055-7

Crossref Full Text | Google Scholar

Ho, Y.-C., Cheung, M.-C., and Chan, A. S. (2003). Music training improves verbal but not visual memory: cross-sectional and longitudinal explorations in children. Neuropsychology 17, 439–50. doi: 10.1037/0894-4105.17.3.439

PubMed Abstract | Crossref Full Text | Google Scholar

Høffding, S. (2014). What is skilled coping? Experts on expertise. J. Conscious. Stud. 21, 49–73.

Google Scholar

Hoppe, C., Splittstößer, C., Fliessbach, K., Trautner, P., Elger, C. E., Weber, B., et al. (2014). Silent music reading: auditory imagery and visuotonal modality transfer in singers and non-singers. Brain Cogn. 91, 35–44. doi: 10.1016/j.bandc.2014.08.002

PubMed Abstract | Crossref Full Text | Google Scholar

Jäncke, L. (2012). The dynamic audio-motor system in pianists. Ann. N. Y. Acad. Sci. 1252, 246–252. doi: 10.1111/j.1749-6632.2011.06416.x

PubMed Abstract | Crossref Full Text | Google Scholar

Jeon, H.-A., and Friederici, A. D. (2015). Degree of automaticity and the prefrontal cortex. Trends Cogn. Sci. 19, 244–250. doi: 10.1016/j.tics.2015.03.003

PubMed Abstract | Crossref Full Text | Google Scholar

Kandel, E. R., Dudai, Y., and Mayford, M. R. (2014). The molecular and systems biology of memory. Cell 157, 163–186. doi: 10.1016/j.cell.2014.03.001

PubMed Abstract | Crossref Full Text | Google Scholar

Karagiorgis, A. T., Chalas, N., Karagianni, M., Papadelis, G., Vivas, A. B., Bamidis, P., et al. (2021). Computerized music-reading intervention improves resistance to unisensory distraction within a multisensory task, in young and older adults. Front. Hum. Neurosci. 15:742607. doi: 10.3389/fnhum.2021.742607

PubMed Abstract | Crossref Full Text | Google Scholar

Kawasaki, A., and Hayashi, N. (2022). Musical instrumental reading affects middle cerebral blood flow and cognitive function. Front. Physiol. 13:966969. doi: 10.3389/fphys.2022.966969

PubMed Abstract | Crossref Full Text | Google Scholar

Kelly, A. M. C., and Garavan, H. (2005). Human functional neuroimaging of brain changes associated with practice. Cerebral Cortex 15, 1089–1102. doi: 10.1093/cercor/bhi005

PubMed Abstract | Crossref Full Text | Google Scholar

Koshimori, Y. (2018). “Neurochemical responses to music,” in The Oxford Handbook of Music and the Brain, eds M. H. Thaut, and D. A. Hodges (Oxford: University of Oxford Press), 332–363. doi: 10.1093/oxfordhb/9780198804123.013.14

Crossref Full Text | Google Scholar

Lee, H.-Y., and Lei, S.-F. (2012). Musical training effect on reading musical notation: evidence from event-related potentials. Percept. Mot. Skills 115, 7–17. doi: 10.2466/22.11.24.PMS.115.4.7-17

PubMed Abstract | Crossref Full Text | Google Scholar

Lee, H.-Y., and Wang, Y.-S. (2011). Visual processing of music notation: a study of event-related potentials. Percept. Mot. Skills 112, 525–535. doi: 10.2466/11.22.24.27.PMS.112.2.525-535

PubMed Abstract | Crossref Full Text | Google Scholar

Lee, Y.-S., Lu, M.-J., and Ko, H.-P. (2007). Effects of skill training on working memory capacity. Learn. Instruct. 17, 336–344. doi: 10.1016/j.learninstruc.2007.02.010

Crossref Full Text | Google Scholar

Lewis, P. A., and Miall, R. C. (2003). Distinct systems for automatic and cognitively controlled time measurement: evidence from neuroimaging. Curr. Opin. Neurobiol. 13, 250–5. doi: 10.1016/S0959-4388(03)00036-9

PubMed Abstract | Crossref Full Text | Google Scholar

Liao, Y.-C., Yang, C.-J., Yu, H.-Y., Huang, C.-J., Hong, T.-Y., Li, W.-C., et al. (2024a). Inner sense of rhythm: percussionist brain activity during rhythmic encoding and synchronization. Front. Neurosci. 18:1342326. doi: 10.3389/fnins.2024.1342326

PubMed Abstract | Crossref Full Text | Google Scholar

Liao, Y.-C., Yang, C.-J., Yu, H.-Y., Huang, C.-J., Hong, T.-Y., Li, W.-C., et al. (2024b). The rhythmic mind: brain functions of percussionists in improvisation. Front. Hum. Neurosci. 18:1418727. doi: 10.3389/fnhum.2024.1418727

PubMed Abstract | Crossref Full Text | Google Scholar

Logan, G. D. (1998). Toward an instance theory of automatization. Psychol. Rev. 95, 492–527. doi: 10.1037/0033-295X.95.4.492

Crossref Full Text | Google Scholar

Lörch, L., Lemaire, B., and Portrat, S. (2023). A hebbian model to account for musical expertise differences in a working memory task. Cognit. Comput. 15, 1620–1639. doi: 10.1007/s12559-023-10138-3

Crossref Full Text | Google Scholar

Lu, C.-I., Greenwald, M., Lin, Y.-Y., and Bowyer, S. M. (2019). Musical transposing versus sight-reading: mapping brain activation with magnetoencephalography. Psychol. Music 49, 581–599. doi: 10.1177/0305735619883692

Crossref Full Text | Google Scholar

Lu, C.-I., Greenwald, M., Lin, Y.-Y., and Bowyer, S. M. (2022). Music, math., and working memory: magnetoencephalography mapping of brain activation in musicians. Front. Hum. Neurosci. 16:866256. doi: 10.3389/fnhum.2022.866256

PubMed Abstract | Crossref Full Text | Google Scholar

Maurer, D., Le Grand, R., and Mondloch, C. J. (2002). The many faces of configural processing. Trends Cogn. Sci. 6, 255–260. doi: 10.1016/S1364-6613(02)01903-4

PubMed Abstract | Crossref Full Text | Google Scholar

Meister, I. G., Krings, T., Foltys, H., Boroojerdi, B., Müller, M., Töpper, R. F., et al. (2004). Playing piano in the mind—An fMRI study on music immagery and performance in pianists. Cogn. Brain Res. 19, 219–228. doi: 10.1016/j.cogbrainres.2003.12.005

Crossref Full Text | Google Scholar

Miller, E. K., and Cohen, J. D. (2001). An integrative theory of prefrontal cortex function. Annu. Rev. Neurosci. 24, 167–202. doi: 10.1146/annurev.neuro.24.1.167

PubMed Abstract | Crossref Full Text | Google Scholar

Mongelli, V., Dehaene, S., Vinckier, F., Peretz, I., Bartolomeo, P., Cohen, L., et al. (2017). Music and words in the visual cortex: the impact of musical expertise. Cortex 86, 260–274. doi: 10.1016/j.cortex.2016.05.016

PubMed Abstract | Crossref Full Text | Google Scholar

Nakada, T., Fujii, Y., Suzuki, K., and Kwee, I. L. (1998). 'Musical Brain' revealed by high-field (3 tesla) functional MRI. NeuroReport 9, 3853–3856. doi: 10.1097/00001756-199812010-00016

PubMed Abstract | Crossref Full Text | Google Scholar

Nichols, E. S., and Grahn, J. A. (2016). Neural correlates of audiovisual integration in music reading. Neuropsychologia 91, 199–210. doi: 10.1016/j.neuropsychologia.2016.08.011

PubMed Abstract | Crossref Full Text | Google Scholar

Olivers, C. N. L., and Eimer, M. (2010). On the difference between working memory and attentional set. Neuropsychologia 49, 1553–1558. doi: 10.1016/j.neuropsychologia.2010.11.033

PubMed Abstract | Crossref Full Text | Google Scholar

Olivers, C. N. L., and Meeter, M. (2008). A boost and bounce theory of temporal attention. Psychol. Rev. Am. Psychol. Assoc. 115, 836–863. doi: 10.1037/a0013395

PubMed Abstract | Crossref Full Text | Google Scholar

Palmer, C., and Drake, C. (1997). Monitoring and planning capacities in the acquisition of music performance skills. Can. J. Exp. Psychol. 51, 369–384. doi: 10.1037/1196-1961.51.4.369

PubMed Abstract | Crossref Full Text | Google Scholar

Palmer, C., and Meyer, R. K. (2000). Conceptual and motor learning in performance. Psychol. Sci. 11, 63–68. doi: 10.1111/1467-9280.00216

Crossref Full Text | Google Scholar

Palmer, C., and van de Sande, C. (1993). Units of knowledge in music performance. J. Exp. Psychol. Learn. Memory Cogn. 19, 457–470. doi: 10.1037//0278-7393.19.2.457

PubMed Abstract | Crossref Full Text | Google Scholar

Pan, Y., Frisson, S., and Jensen, O. (2021). Neural evidence for lexical parafoveal processing. Nat. Commun. 12:5234. doi: 10.1038/s41467-021-25571-x

PubMed Abstract | Crossref Full Text | Google Scholar

Paraskevopoulos, E., Kuchenbuch, A., Herholz, S. C., Foroglou, N., Bamidis, P., Pantev, C., et al. (2014b). Tones and numbers: a combined EEG-MEG study on the effects of musical expertise in magnitude comparisons of audiovisual stimuli. Hum. Brain Mapp. 35, 5389–5400. doi: 10.1002/hbm.22558

PubMed Abstract | Crossref Full Text | Google Scholar

Paraskevopoulos, E., Kuchenbuch, A., Herholz, S. C., and Pantev, C. (2014a). Multisensory Integration during short-term music reading training enhances both uni- and multisensory cortical processing. J. Cogn. Neurosci. 26, 2224–2238. doi: 10.1162/jocn_a_00620

PubMed Abstract | Crossref Full Text | Google Scholar

Penttinen, M., Huovinen, E., and Ylitalo, A.-K. (2014). Reading ahead: adult music students' eye movements in temporally controlled performances of a children's song. Int. Soc. Music Educ. 33, 36–50. doi: 10.1177/0255761413515813

Crossref Full Text | Google Scholar

Persson, R. (1996). Brilliant performers as teachers: a case study of commonsense teaching in a conservatoire setting. Int. J. Music Educ. 28, 25–36. doi: 10.1177/025576149602800103

Crossref Full Text | Google Scholar

Porter, R., and Lemon, R. (1995). Corticospinal Function and Voluntary Movement. Oxford: Oxford University Press. doi: 10.1093/acprof:oso/9780198523758.001.0001

Crossref Full Text | Google Scholar

Prinz, W. (1997). Perception and action planning. Eur. J. Cogn. Psychol. 9, 129–154. doi: 10.1080/713752551

Crossref Full Text | Google Scholar

Proverbio, A. M., Arcuri, G., Pantaleo, M. M., Zani, A., and Manfredi, M. (2024). The key role of the right posterior fusiform gyrus in music reading: an electrical neuroimaging study on 90 readers. Front. Cogn. 3:1323220. doi: 10.3389/fcogn.2024.1323220

Crossref Full Text | Google Scholar

Proverbio, A. M., and Sanoubari, E. (2024). Music literacy shapes the specialization of a right hemispheric word reading area. Neuroimage Rep. 4:100219. doi: 10.1016/j.ynirp.2024.100219

PubMed Abstract | Crossref Full Text | Google Scholar

Proverbio, A. M., and Valtolina, M. (2025). Musical expertise modulates embodied processing of biological motion and audiovisual-motor integration in rhythmic hand tapping. NeuroImage 315:121287. doi: 10.1016/j.neuroimage.2025.121287

PubMed Abstract | Crossref Full Text | Google Scholar

Reybrouck, M. (1996). “Gestalt concepts and music: limitations and possibilities,” in Joint International Conference on Cognitive and Systematic Musicology (Berlin, Heidelberg: Springer Berlin Heidelberg), 57–69. doi: 10.1007/BFb0034107

Crossref Full Text | Google Scholar

Ross, V., Murat, Z. H., Buniyamin, N., and Mohd-Zain, Z. (2013). “Violinists playing with and without music notation: investigating hemispheric brainwave activity,” in Intelligent Systems for Science and Information: Extended and Selected Results from the Science and Information Conference (Cham: Springer International Publishing), 153–169. doi: 10.1007/978-3-319-04702-7_9

Crossref Full Text | Google Scholar

Ross, V., Murat, Z. H., Buniyamin, N., and Mohd-Zain, Z. (2014). “Violinists playing with and without music notation: investigating hemispheric brainwave activity,” in Intelligent Systems for Science and Information. Studies in Computational Intelligence, Vol. 542, eds. L. Chen, S. Kapoor, and R. Bhatia (Cham: Springer). doi: 10.1007/978-3-319-04702-7_9

Crossref Full Text | Google Scholar

Sachs, M., Kaplan, J., Sarkissian, A. D., and Habibi, A. (2017). Increased engagement of the cognitive control network associated with music training in children during an fMRI Stroop task. PloS ONE 12:e0187254. doi: 10.1371/journal.pone.0187254

PubMed Abstract | Crossref Full Text | Google Scholar

Sala, G., and Gobet, F. (2017). Does far transfer exist? Negative evidence from chess, music., and working memory training. Assoc. Psychol. Sci. 26, 515–520. doi: 10.1177/0963721417712760

PubMed Abstract | Crossref Full Text | Google Scholar

Schacter, D. L., Addis, D. R., and Buckner, R. L. (2007). Remembering the past to imagine the future: the prospective brain. Nat. Rev. 8, 657–661. doi: 10.1038/nrn2213

PubMed Abstract | Crossref Full Text | Google Scholar

Schacter, D. L., Benoit, R. G., and Szpunar, K. K. (2017). Episodic future thinking: mechanisms and functions. Curr. Opin. Behav. Sci. 17, 41–50. doi: 10.1016/j.cobeha.2017.06.002

PubMed Abstract | Crossref Full Text | Google Scholar

Schneider, W., and Shiffrin, R. M. (1977). Controlled and automatic human information processing: I. Detection, search., and attention. Psychol. Rev. 84, 1–66. doi: 10.1037/0033-295X.84.1.1

Crossref Full Text | Google Scholar

Schön, D., Anton, J. L., Roth, M., and Besson, M. (2002). An fMRI study of music sight-reading. Neuroreport 13, 2285–2289. doi: 10.1097/00001756-200212030-00023

PubMed Abstract | Crossref Full Text | Google Scholar

Schön, D., and Besson, M. (2002). Processing pitch and duration in music reading: a RT–ERP study. Neuropsychologia 40, 868–878. doi: 10.1016/S0028-3932(01)00170-1

PubMed Abstract | Crossref Full Text | Google Scholar

Schubert, E. (2013). Emotion felt by the listener and expressed by the music: literature review and theoretical perspectives. Front. Psychol. 4:837. doi: 10.3389/fpsyg.2013.00837

PubMed Abstract | Crossref Full Text | Google Scholar

Schulze, K., and Koelsch, S. (2012). Working memory for speech and music. Ann. N. Y. Acad. Sci. 1252, 229–326. doi: 10.1111/j.1749-6632.2012.06447.x

PubMed Abstract | Crossref Full Text | Google Scholar

Scott, S. H. (2004). Optimal feedback control and the neural basis of volitional motor control. Nat. Rev. Neurosci. 5, 532–545. doi: 10.1038/nrn1427

PubMed Abstract | Crossref Full Text | Google Scholar

Sergent, J., Zuck, E., Terriah, S., and MacDonald, B. (1992). Distributed neural network underlying musical sight-reading and keyboard performance. Science 257, 106–109. doi: 10.1126/science.1621084

PubMed Abstract | Crossref Full Text | Google Scholar

Servant, M., Cassey, P., Woodman, G. F., and Logan, G. D. (2018). Neural bases of automaticity. J. Exp. Psychol. Learn. Mem. Cogn. 44, 440–464. doi: 10.1037/xlm0000454

PubMed Abstract | Crossref Full Text | Google Scholar

Shenoy, K. V., Sahani, M., and Churchland, M. M. (2013). Cortical control of arm movements: a dynamical systems perspective. Annu. Rev. Neurosci. 8, 337–359. doi: 10.1146/annurev-neuro-062111-150509

PubMed Abstract | Crossref Full Text | Google Scholar

Sidnell, R. G. (1986). Motor learning in music education. Psychomusicology 6, 7–18. doi: 10.1037/h0094198

Crossref Full Text | Google Scholar

Silverman, M. J. (2010). The effect of pitch, rhythm., and familiarity on working memory and anxiety as measured by digit recall performance. J. Music Ther. 47, 70–83. doi: 10.1093/jmt/47.1.70

PubMed Abstract | Crossref Full Text | Google Scholar

Simoens, V. L., and Tervaniemi, M. (2013). Auditory short-term memory activation during score reading. PLoS ONE 8:e53691. doi: 10.1371/journal.pone.0053691

PubMed Abstract | Crossref Full Text | Google Scholar

Sloboda, J. (2004). Exploring the Musical Mind: Cognition, Emotion, Ability, Function. Oxford: Oxford University Press. doi: 10.1093/acprof:oso/9780198530121.001.0001

Crossref Full Text | Google Scholar

Sluming, V., Brooks, J., Howard, M., Downes, J. J., and Roberts, N. (2007). Broca's Area Supports enhanced visuospatial cognition in orchestral musicians. J. Neurosci. 27, 3799–3806. doi: 10.1523/JNEUROSCI.0147-07.2007

PubMed Abstract | Crossref Full Text | Google Scholar

Stambaugh, L. A. (2011). When repetition isn't the best practice strategy: effects of blocked and random practice schedules. J. Res. Music Educ. 58, 368–383. doi: 10.1177/0022429410385945

Crossref Full Text | Google Scholar

Stewart, L. (2005a). A neurocognitive approach to music reading. Ann. N. Y. Acad. Sci. 106, 377–386. doi: 10.1196/annals.1360.032

PubMed Abstract | Crossref Full Text | Google Scholar

Stewart, L. (2005b). Neurocognitive studies of musical literacy acquisition. Musicae Sci. 9, 223–237. doi: 10.1177/102986490500900204

Crossref Full Text | Google Scholar

Stewart, L., Henson, R., Kampe, K., Walsh, V., Turner, R., Firth, U., et al. (2003a). Becoming a pianist: an fMRI study of musical literacy acquisition. Ann. N. Y. Acad. Sci. 999, 204–208. doi: 10.1196/annals.1284.030

PubMed Abstract | Crossref Full Text | Google Scholar

Stewart, L., Henson, R., Kampe, K., Walsh, V., Turner, R., Frith, U., et al. (2003b). Brain changes after learning to read and play music. Neuroimage 20, 71–83. doi: 10.1016/S1053-8119(03)00248-9

PubMed Abstract | Crossref Full Text | Google Scholar

Strobach, T., Wendt, M., and Janczyk, M. (2018). Multitasking: Executive Functioning in Dual-Task and Task-Switching Situations. Lausanne: Frontiers Media SA. doi: 10.3389/978-2-88945-453-2

Crossref Full Text | Google Scholar

Sweller, J. (2011). “Cognitive load theory,” in The Psychology of Learning and Motivation: Cognition in Education, eds J. P. Mestre, and B. H. Ross (San Diego, CA: Elsevier Academic Press), 37–76. doi: 10.1007/978-1-4419-8126-4

Crossref Full Text | Google Scholar

Tan, J., Di Bernard Luft, C., and Bhattacharya, J. (2024). The after-glow of flow: neural correlates of flow in musicians. J. Soc. Neurosci. Creat. 36, 469–490. doi: 10.1080/10400419.2023.2277042

Crossref Full Text | Google Scholar

Tenney, J., and Polansky, L. (1980). Temporal gestalt perception in music. J. Music Theory 24, 205–241. doi: 10.2307/843503

Crossref Full Text | Google Scholar

Terhardt, E., Yost, W. A., and Watson, C. S. (1987). “Gestalt principles and music perception,” in Auditory Processing of Complex Sounds, 1st edn, eds. W. A. Yost and C. S. Watson (New York, NY: Routledge), 157–166.

Google Scholar

Thaut, M. H., McIntosh, G. C., and Hoemberg, V. (2015). Neurobiological foundations of neurologic music therapy: rhythmic entrainment and the motor system. Front. Psychol. 18:1185. doi: 10.3389/fpsyg.2014.01185

PubMed Abstract | Crossref Full Text | Google Scholar

Tsao, A., Aryana Yousefzadeh, S., Meck, W. H., Moser, M.-B., and Moser, E. I. (2022). The Neural bases for timing of durations. Nature 23, 646–665. doi: 10.1038/s41583-022-00623-3

PubMed Abstract | Crossref Full Text | Google Scholar

Tulving, E. (2002). Episodic memory: from mind to brain. Annu. Rev. Psychol. 53, 1–25. doi: 10.1146/annurev.psych.53.100901.135114

PubMed Abstract | Crossref Full Text | Google Scholar

Vul, E., Harris, C., Winkielman, P., and Pashler, H. (2009). Puzzlingly high correlations in fmri studies of emotion, personality., and social cognition. Perspect. Psychol. Sci. 4, 274–290. doi: 10.1111/j.1745-6924.2009.01125.x

PubMed Abstract | Crossref Full Text | Google Scholar

Wakita, M. (2016). Interaction between perceived action and music sequences in the left prefrontal area. Front. Hum. Neurosci. 10:656. doi: 10.3389/fnhum.2016.00656

PubMed Abstract | Crossref Full Text | Google Scholar

Waters, A. J., Townsend, E., and Underwood, G. (1998). Expertise in musical sightreading: a study of pianists. Br. J. Psychol. 89, 123–149. doi: 10.1111/j.2044-8295.1998.tb02676.x

Crossref Full Text | Google Scholar

Wertheimer, M. (1912). Experimentelle Studien über das Sehen von Bewegung. Zeitschrift für Psychologie 61, 161–265.

Google Scholar

Wieth, M. B., and Burns, B. D. (2014). Rewarding multitasking: negative effects of an incentive on problem solving under divided attention. J. Problem Solving, 7, 60–73. doi: 10.7771/1932-6246.1163

Crossref Full Text | Google Scholar

Wilkins, N. J., and Rawson, K. A. (2011). Controlling retrieval during practice: implications for memory-based theories of automaticity. J. Mem. Lang. 65, 208–211. doi: 10.1016/j.jml.2011.03.006

Crossref Full Text | Google Scholar

Wilson, F. R. (1986). Tone Deaf & All Thumbs? An Invitation to Music-Making for Late Bloomers and Non-Prodigies. New York, NY: Vintage Books.

Google Scholar

Wolff, A., Berberian, N., Golesorkhi, M., Gomez-Pilar, J., Zilio, F., Northoff, G., et al. (2022). Intrinsic neural timescales: temporal integration and segregation. Trends Cogn. Sci. 26, 159–173. doi: 10.1016/j.tics.2021.11.007

PubMed Abstract | Crossref Full Text | Google Scholar

Wöllner, C., and Williamon, A. (2007). An exploratory study of the role of performance feedback and musical imagery in piano playing. Res. Stud. Music Educ. 29, 39–54. doi: 10.1177/1321103X07087567

Crossref Full Text | Google Scholar

Wolpert, D. M., Chris Miall, R., and Kawato, M. (1998). Internal models in the cerebellum. Trends Cogn. Sci. 2, 338–347. doi: 10.1016/S1364-6613(98)01221-2

Crossref Full Text | Google Scholar

Wong, Y. K., and Gauthier, I. (2010a). Holistic processing of musical notation: dissociating failures of selective attention in experts and novices. Cogn. Affect. Behav. Neurosci. 10, 541–551. doi: 10.3758/CABN.10.4.541

PubMed Abstract | Crossref Full Text | Google Scholar

Wong, Y. K., and Gauthier, I. (2010b). A multimodal neural network recruited by expertise with musical notation. J. Cogn. Neurosci. 22, 695–713. doi: 10.1162/jocn.2009.21229

PubMed Abstract | Crossref Full Text | Google Scholar

Yeşil, B., and Nal, S. (2017). An investigation on the effects of music training on attention and working memory in adults. Anat. J. Psychiatr. 18, 531–535. doi: 10.5455/apd.259201

Crossref Full Text | Google Scholar

Yurgil, K. A., Velasquez, M. A., Winston, J. L., Reichman, N. B., and Colombo, P. J. (2020). Music training, working memory., and neural oscillations: a review. Front. Psychol. 11, 1–17. doi: 10.3389/fpsyg.2020.00266

PubMed Abstract | Crossref Full Text | Google Scholar

Zatorre, R. J. (2012). Beyond auditory cortex: working with musical thoughts. Ann. N. Y. Acad. Sci. 1252, 222–228. doi: 10.1111/j.1749-6632.2011.06437.x

PubMed Abstract | Crossref Full Text | Google Scholar

Zatorre, R. J., Chen, J. L., and Penhune, V. B. (2007). When the brain plays music: auditory–motor interactions in music perception and production. Nat. Rev. Neurosci. 8, 547–558. doi: 10.1038/nrn2152

PubMed Abstract | Crossref Full Text | Google Scholar

Zimmerman, E., and Lahav, A. (2012). The multisensory brain and its ability to learn music. Ann. N. Y. Acad. Sci. 1252, 179–184. doi: 10.1111/j.1749-6632.2012.06455.x

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: cognition, dual processing, memory, music notation, music reading, performance, prospective planning

Citation: Heath KL (2026) Bi-temporal processing in music notation reading: a theory linking prediction, memory, and automaticity. Front. Cognit. 4:1689600. doi: 10.3389/fcogn.2025.1689600

Received: 20 August 2025; Revised: 09 December 2025;
Accepted: 29 December 2025; Published: 06 February 2026.

Edited by:

Alice Mado Proverbio, University of Milano-Bicocca, Italy

Reviewed by:

Gabriel Byczynski, University of Geneva, Switzerland
Yin-Chun Liao, Taipei National University of the Arts, Taiwan
Damien Sagrillo, University of Luxembourg, Luxembourg

Copyright © 2026 Heath. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Karen L. Heath, a2FyZW4uaGVhdGhAdW5pbWVsYi5lZHUuYXU=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.