Original Research ARTICLE
An fMRI comparison of neural activity associated with recognition of familiar melodies in younger and older adults
- 1Centre for Neuroscience Studies, Queen's University, Kingston, ON, Canada
- 2Music Cognition Lab, Department of Psychology, Queen's University, Kingston, ON, Canada
- 3Cognitive Neuroscience of Communication and Hearing, Department of Psychology, Queen's University, Kingston, ON, Canada
Several studies of semantic memory in non-musical domains involving recognition of items from long-term memory have shown an age-related shift from the medial temporal lobe structures to the frontal lobe. However, the effects of aging on musical semantic memory remain unexamined. We compared activation associated with recognition of familiar melodies in younger and older adults. Recognition follows successful retrieval from the musical lexicon that comprises a lifetime of learned musical phrases. We used the sparse-sampling technique in fMRI to determine the neural correlates of melody recognition by comparing activation when listening to familiar vs. unfamiliar melodies, and to identify age differences. Recognition-related cortical activation was detected in the right superior temporal, bilateral inferior and superior frontal, left middle orbitofrontal, bilateral precentral, and left supramarginal gyri. Region-of-interest analysis showed greater activation for younger adults in the left superior temporal gyrus and for older adults in the left superior frontal, left angular, and bilateral superior parietal regions. Our study provides powerful evidence for these musical memory networks due to a large sample (N = 40) that includes older adults. This study is the first to investigate the neural basis of melody recognition in older adults and to compare the findings to younger adults.
Aging affects brain mechanisms underlying many cognitive processes including those related to memory. Even when younger and older adults perform equally well on a cognitive task, older adults may display increased or decreased neural activity in specific brain regions, depending on the actual task (Grady, 2012). Aging-related changes in memory, whether for performance or for the corresponding neural activity, may vary depending on many factors, such as type of memory (e.g., semantic, episodic), memory process (e.g., encoding, retrieval), sensory modality involved in the memory, age of the individual, and possibly age of the memory itself. Although studies of musical memory are scarce compared to those with verbal or pictorial stimuli, they suggest that musical memory may be preserved in aging and even in Alzheimer's disease (Vanstone et al., 2012; Kerer et al., 2013). However, studies of brain activity associated with musical memory have largely limited their participants to younger adults. Thus, the effects of aging on neural activity related to musical memory remain unknown.
Though, not all memories can be neatly classified into any single category, a distinction between semantic and episodic memory, as proposed by Tulving (1972), is supported by behavioral and neural evidence (Mayes and Montaldi, 2001; Tulving, 2002; Winocur and Moscovitch, 2011; Yee et al., 2013). Semantic memory is knowledge about the world, including concrete objects and abstract concepts, as well as relationships between them, whereas episodic memory involves details about the temporal and spatial context of an event. Musical memory, that is, memory of the music itself, has been considered a type of semantic memory since it comprises knowledge of the musical characteristics of a tune (i.e., the way it sounds, and not any “meta” information such as composer or title). Several studies have regarded musical semantic memory as the underlying basis of familiar melody recognition (Halpern and Zatorre, 1999; Platel et al., 2003; Schulkind, 2004; Groussard et al., 2009; Hailstone et al., 2009; Omar et al., 2010; Weinstein et al., 2011; Vanstone et al., 2012).
The ability to recognize a melody in as little as 2 s, with only 3–6 notes (Bella et al., 2003) requires fast retrieval from semantic memory. A musical lexicon, akin to the concept of a language-based lexicon, is thought to be a type of music storage system that is accessed effortlessly during recognition of familiar music (Peretz, 1996). This lexicon is assumed to be a module in a postulated system of hierarchically organized, independently functioning modules involved in the processing of music in the brain (Peretz and Coltheart, 2003). “The musical lexicon is a representational system that contains all the representations of the specific musical phrases to which one has been exposed during one's lifetime” (Peretz and Coltheart, 2003, p. 690). When a match for the musical phrase under consideration is found in the musical lexicon, what follows is recognition in the form of a feeling of familiarity and/or possible recall of associated information such as song title or composer, or autobiographical details from one's past life. In this study, we associate the musical lexicon with musical semantic memory. Double dissociations found in studies of deficits in patients with brain damage or congenital disorders suggest that recognition of music involves at least some brain areas that are distinct from those involved in recognition of other types of auditory input (Peretz et al., 1994; Peretz, 1996, 2001; Ayotte et al., 2000).
In an fMRI study, Peretz et al. (2009) found that listening to familiar music, compared to unfamiliar music, correlated with activity in the right superior temporal area. A voxel-based morphometry analysis of patients with semantic dementia also found that the amount of atrophy in this region, especially in the right superior temporal pole, was associated with a reduced ability to recognize famous tunes (Hsieh et al., 2011). When using a familiar melody pitch error detection task to test melody recognition in a combined group of patients with frontotemporal dementia, semantic dementia, and Alzheimer's disease, as well as healthy controls, Johnson et al. (2011) found that recognition correlated with gray matter volume of right-sided brain areas in the inferior frontal gyrus, inferior and superior temporal gyri, and the temporal pole. Interestingly, performance in this task was not correlated with any neuropsychological tests, whereas recall of familiar melody titles did correlate with tests requiring semantic knowledge. This suggests that the title recall may be related to non-musical semantic processing (due to the observed correlation with standard neuropsychological tests of semantic memory), whereas the familiar melody pitch error detection task may be exclusively testing the musical aspects of melody recognition (which were not part of the standard tests).
Groussard et al. (2010), when comparing musical against verbal familiarity in an fMRI study, similarly identified increased activation in the superior temporal gyrus. In the same study, activation within a large network, including the inferior frontal, posterior inferior and middle temporal, medial superior frontal gyri, and the right superior temporal pole, was associated with the level of musical familiarity. The authors suggest that the left superior temporal gyrus may be involved in access to, and the left inferior frontal area in selection from, musical semantic memory, whereas the right superior temporal gyrus activation may reflect retrieval of perception-based musical memory. In a study of familiarity with music and odor stimuli, Plailly et al. (2007) found a largely left-sided activation related to music familiar to the participant (compared to unfamiliar) in several regions of the frontal gyrus, superior temporal sulcus, cingulate, supramarginal, precuneus, mid-occipital, and right angular gyri. Certain regions of activation were common to both olfactory and musical familiarity, but the superior temporal sulcus was among those exclusive to musical familiarity1.
None of the aforementioned studies on the neural basis of tune recognition were conducted with older adults. In non-musical domains, neural activation for memories is thought to undergo an age-related shift from medial temporal lobe structures such as the hippocampus and entorhinal cortex to neocortical areas such as the prefrontal cortex (Haist et al., 2001; Douville et al., 2005; Frankland and Bontempi, 2005; Smith and Squire, 2009; Squire and Wixted, 2011). A similar shift is also possible for musical memory. The goal of the present study was to determine if this age-based difference might also occur in musical memory. To this effect, we used fMRI to investigate the neural correlates of melody recognition in younger and older adults. In order to isolate this process as much as possible, we contrasted the brain activity while listening to familiar melodies against listening to unfamiliar melodies, where the unfamiliar melodies had many of the same musical properties as the familiar melodies, yet remained unrecognizable. This paradigm was also used by Peretz et al. (2009). Recognition of familiar melodies across groups was expected to be associated with predominant activation in the superior and middle temporal as well as inferior frontal regions; for the superior temporal area, we expected a stronger activation on the right. Based on evidence of the effects of aging on the neural correlates of semantic memory in non-musical domains, we expected older adults (relative to younger adults) to have increased activity in the prefrontal cortex, specifically the inferior frontal gyrus, and reduced activity in the superior temporal region and in medial temporal lobe structures, i.e., the hippocampus and surrounding areas. A reduced dependence on the medial temporal lobe would provide a plausible neuronal basis for preserved musical memories in conditions where function in this region is compromised, such as starting early in Alzheimer's disease.
Materials and Methods
All participants were female, right-handed, and non-musicians, defined for our purposes as those with a maximum of 3 years of music training. Younger (n = 20; age range = 18–25 years, mean = 20) and older (n = 20; age range = 65–84 years, mean = 71) adults did not differ in years of education [M = 14.1 and M = 14.7, respectively, t(38) = 1.1, p = 0.29]. Older participants scored in the normal range on the Mini-Mental State Examination (Folstein et al., 1975; M = 29.5, SD = 0.87, range = 27–30).
The familiar and unfamiliar tunes had been tested for familiarity in a pilot study. On a familiarity scale from 0 (“not-at-all familiar”) to 3 (“definitely familiar”), where the level of familiarity indicated the certainty of the participant having heard the tune anywhere during their lives, a cut-off of 1.5 for familiar vs. unfamiliar tunes was used for younger and older participants who met the same criteria as in this study.
Familiar melodies were excerpts from well-known instrumental pieces without associated lyrics in order to avoid stimulating brain activity specifically related to language use. Unfamiliar melodies were reversed versions of familiar melodies, i.e., with reversed note order and preserved tempo, thus largely matching the tonal and temporal qualities of their original counterparts in order to limit the difference between the two versions to familiarity as much as possible (Hébert et al., 1995). Minor corrections were made to the unfamiliar melodies, when required, due to any irregularities in the metrical structure2. It was important to maximally control for perceptual differences between the familiar and unfamiliar melodies while ensuring that the unfamiliar melodies were unrecognizable. Figure 1 shows a sample melody and its reversed version. A complete list of familiar melodies used in this study is given in the Appendix. The selected familiar and unfamiliar melodies were similar in style to those used by Peretz et al. (2009).
Figure 1. An original (familiar) melody and its reversed (unfamiliar) version. The order of the notes in the familiar melody, an excerpt from Beethoven's “Symphony No. 5” (top), is reversed to create an unfamiliar melody (bottom).
Four types of stimuli were used during the functional imaging: familiar melodies, unfamiliar melodies, signal-correlated noise (SCN), and silent trials. Each stimulus was 8.5 s in length, including a 500 ms fade-out at the end of the melodies. The melodies were edited using Sibelius3 v6 music notation software, and exported as wave files using Kontakt Player 2 by Native Instruments4 running the Steinway Grand Piano sample bank by Garritan Personal Orchestra5 v3. The sound files were encoded with 16 bits per sample at a rate of 44.1 kHz using Audition v3 by Adobe6. The SCN stimuli were created with Praat software (http://www.praat.org) from randomly selected unfamiliar melodies. SCN preserves the amplitude envelope and spectral profile of a waveform (Davis and Johnsrude, 2003), but does not contain spectral details of the sound. SCN was used as a reference stimulus for the melodies since it retains some of the temporal patterns and pitch from the original melody while remaining unrecognizable.
Each participant was tested over two sessions on separate days. The main goal of the first session was to familiarize participants with the experimental setup by using a sham magnetic resonance (MR) system7 to simulate a block of the functional imaging protocol. Participants also completed demographic, health, and music-related questionnaires and cognitive tests (older adults only) during the first session. All scanning occurred during the second session. A total of 120 stimuli (39 familiar tunes, 39 unfamiliar tunes, 21 signal-correlated noise, and 21 silent trials) were presented in a pseudorandomized order, divided into four blocks containing 30 stimuli each, with roughly the same numbers of each stimulus type in each block. The participants were asked to simply listen to the stimuli, without providing any active responses, in order to avoid activation related to decision-making processes or motor activity related to button presses. After the scanning session, the participants listened to the same melodies as in the scanner to rate their level of familiarity for each tune. Participants were instructed to base the rating on general lifetime familiarity, and not on whether the tune had been heard in the scanner. The rating scale ranged from 0 (“not-at-all familiar”) to 3 (“definitely familiar”)8.
A Siemens 3 Tesla MAGNETOM Trio whole body MRI scanner with a 12-channel head matrix coil was used for the scanning. The auditory stimuli were delivered through a NordicNeuroLab Audio System with headphones that also provided a noise attenuation of approximately 30 dB9. Detailed structural images were obtained at the start of scanning using a T1-weighted magnetization-prepared rapid gradient echo (MP-RAGE) single-shot sequence to acquire 176 sagittal slices with a field of view (FOV) = 256 × 256 mm, in-plane resolution = 1.0 mm × 1.0 mm, slice thickness = 1.0 mm, flip angle (FA) = 9°, repetition time (TR) = 1760 ms, echo time (TE) = 2.2 ms. This scan took 7.5 min. After the structural scan, a sample tune was played to allow adjustment of the sound intensity to a level comfortable for the participant listening through the headphones. The sparse-sampling functional imaging protocol was then started. A single brain volume was acquired immediately following each 8.5-s stimulus. Scans occurred every 10.5 s using a T2*-weighted GE-EPI interleaved sequence to acquire 32 axial slices with FOV = 211 × 211 mm, in-plane resolution = 3.3 × 3.3 mm, slice thickness = 3.3 mm with a 25% gap, FA = 78°, acquisition time = 2000 ms, TE = 30 ms, TR = 10.5 s. In this way, sounds were presented in the silent intervals between successive scans, and scans were acquired at a point in time at which the hemodynamic response to the stimulus would be measurable.
During the functional imaging, 120 volumes were obtained over the course of 4 blocks that lasted 5.6 min each.
Imaging Data Analysis
SPM10 software was used to analyze imaging data. For each participant, processing was performed using the following steps, in the given order: (1) DICOM to NIfTI conversion; (2) spatial preprocessing: realignment (motion correction), co-registration of the functional time-series with the T1-weighted structural image, segmentation and normalization of the T1-weighted image to MNI152 space, application of the T1-derived deformation parameters from the normalization step to the fMRI time series; and smoothing using a Gaussian kernel with a FWHM of 8 mm (3) model specification and estimation; and (4) contrast specification and generation of statistical parametric maps. The familiar-SCN and unfamiliar-SCN contrasts were computed for each participant. These contrasts formed the basis of a full-factorial model in SPM using melody familiarity as a within-subject factor with two levels (familiar and unfamiliar) and age as a between-group factor with two levels (younger and older adults). The results pertinent to our study were activation related to the main effect of tune recognition and any age group differences in this activation resulting from a familiarity-age interaction. Whole brain and region-of-interest (ROI) analyses were performed for the resulting contrasts. We used the SPM toolkit MarsBaR (Brett et al., 2002) with automated anatomical labeling (Tzourio-Mazoyer et al., 2002) to examine activation in specific anatomical ROIs. Due to our hypothesis, we tested frontal regions for greater recognition-related activity in older adults compared to younger adults, and conversely, we tested hippocampal and superior temporal regions for greater activity in younger relative to older adults. In order to focus on specific areas, we selected the pre-defined anatomical ROIs based on regions showing significant overall recognition-related activation across all participants. Additionally, a regression analysis examined the effect of age on tune recognition activity separately in each of the younger and older groups. In all cases, a family-wise error-corrected p-threshold (pFWEc) of 0.05 (Worsley et al., 1996) was applied to all statistical probability maps in order to determine significant areas of activation in SPM.
The study was cleared for ethics compliance by the Queen's University Health Sciences and Affiliated Teaching Hospitals Research Ethics Board. All participants gave written informed consent in accordance with this Ethics Board's guidelines and received $30 over two sessions.
There was no effect of age on familiarity ratings of familiar tunes [younger M = 2.73, SD = 0.22; older M = 2.85, SD = 0.23; t(38) = 1.6, p = 0.12] or of unfamiliar tunes [younger M = 0.58, SD = 0.35; older M = 0.61, SD = 0.49; t(34) = 0.17, p = 0.87; Figure 2]. Overall, familiar tunes were rated as more familiar than unfamiliar tunes by all participants [familiar M = 2.79, SD = 0.19; unfamiliar M = 0.59, SD = 0.32; t(62) = 36.1, p < 0.001].
Figure 2. Mean familiarity ratings of familiar and unfamiliar tunes by younger and older adults. Error bars represent standard deviation.
Recognition of a tune was operationalized as a familiarity rating of 2 or 3 for a familiar tune and 0 or 1 for a unfamiliar tune (on a familiarity scale from 0 to 3). Younger and older adults did not differ in percentage of tunes thus recognized [younger M = 88.4, SD = 7.9; older M = 89.8, SD = 9.2, t(38) = 0.5, p = 0.62].
Activation associated with processing of sound was determined in order to verify the accuracy of the design matrix (assignment of conditions to scans) in this study. In the sound-silence contrast, sound comprised all conditions that contained a melody or signal-correlated noise. The individual contrasts were used in a second-level random effects analysis to determine group activation and differences in activation between groups. Whole-brain analysis showed the expected large clusters of activation in the right and left temporal gyri, with respective peaks at MNI x, y, z (in mm) = 49, −24, 10 (t = 18.9, cluster size = 1033 voxels) and MNI x, y, z (in mm) = −44, −20, 7 (t = 17.9, cluster size = 1008 voxels). Younger and older adults did not differ in sound-related activation using whole-brain analysis with a family-wise error corrected p-threshold of 0.05.
For tune recognition in the whole brain analysis, the 2 × 2 mixed ANOVA showed no main effect of age, but the main effect of melody familiarity yielded several cortical regions of activation. The largest recognition-related activation cluster (2304 voxels) appeared in the bilateral inferior frontal, left middle orbitofrontal, and right superior temporal gyri (Figure 3). Other clusters of activation were found within the left supramarginal, bilateral superior frontal, left cingulate, and bilateral precentral gyri (Table 1).
Figure 3. Activation associated with melody familiarity for all participants. pFWEc < 0.05; N = 40. The activation is displayed on the average of all participants' spatially normalized structural images. The right side of each image represents the right side of the brain. Coordinates are in MNI space.
No interaction between melody familiarity and age was found for whole brain analysis, i.e., there were no differences between younger and older adults in recognition-related activation. The regression analysis for the main effect of age on tune recognition also yielded no regions of significant activation in either the younger or the older groups. However, ROI analysis in MarsBaR (Brett et al., 2002) on cortical areas with overall recognition-related activation (Table 1), when tested bilaterally, did show some group differences (Figure 4). Younger adults had greater recognition-related activation than older adults in the left superior temporal gyrus [t(38) = 2.27, p = 0.013], whereas older adults showed greater activation than younger adults in the left superior frontal gyrus [t(38) = 1.81, p = 0.037]. In accordance with our hypothesis of greater recognition-related activation in the medial temporal lobe in younger adults, we tested the hippocampal region. Although the right hippocampus did have greater activation in younger adults, the results were not significant [t(38) = 1.49, p = 0.070]. Lastly, an exploratory analysis on the parietal region showed greater recognition-related activity in older adults in the left angular gyrus [t(38) = 3.35, p < 0.001] and bilateral superior parietal gyrus [t(38) = 1.85, p = 0.034].
Figure 4. Age differences in activation associated with melody familiarity. pFWEc < 0.05; ROI Analysis using MarsBaR. (A) Regions where activation was greater for older adults than younger adults, and (B) regions where activation was greater for younger adults than older adults. For the coronal and axial slices, the right side of the image represents the right side of the brain. Coordinates are in MNI space.
We found robust activation for familiar melodies (compared to unfamiliar melodies) in several frontal lobe areas, along with the superior temporal and supramarginal gyri, when testing the full set of 40 participants from the two age groups conjointly. The specific frontal areas of activity were in the bilateral inferior frontal, left middle orbitofrontal, and bilateral superior frontal gyri. Subcortical regions with recognition-based activation were in the thalamus, putamen, and brainstem. ROI analysis showed greater activation in the left superior temporal gyrus for younger adults and in superior frontal, angular, and superior parietal regions for older adults.
Activation associated with familiar melodies must be understood in the context of the full experience of the participant during the imaging. Recognition of a familiar melody can sometimes be accompanied by recall of a specific incident when the melody may have been heard or of other memories associated with the melody. Since unfamiliar melodies are less likely to evoke comparable memories, the recognition-related activation could include episodic memory or perhaps details about the melody itself, such as the composer. In addition, familiar melodies may be associated with specific emotions. Thus, activation in brain regions that are known to be involved in these functions may be attributed to the corresponding aspects of the participant experience instead of strict access to musical memory itself.
The regions of activation associated with tune recognition in this study have also been identified in various previous studies of musical semantic memory. The inferior frontal and superior temporal areas, which were part of the largest cluster (2304 voxels) of activation, have been frequently reported. The inferior temporal region (Halpern and Zatorre, 1999; Platel et al., 2003; Satoh et al., 2006; Plailly et al., 2007; Groussard et al., 2010) may participate in retrieval from semantic memory in both musical (Groussard et al., 2010) and non-musical (Binder et al., 2009) domains. According to Binder and Desai (2011), the left inferior frontal gyrus activation occurs during selection from semantic memory from among several alternatives.
The role of the superior temporal region has been established in numerous studies of musical semantic memory (Halpern and Zatorre, 1999; Platel, 2005; Satoh et al., 2006; Plailly et al., 2007; Groussard et al., 2009; Janata, 2009; Peretz et al., 2009; Hsieh et al., 2011; Pereira et al., 2011). This region may be unique to the musical domain. When examining non-musical semantic memory in a meta-analysis of 120 fMRI studies, Binder et al. (2009) found little evidence of superior temporal involvement. This region is also involved in processing of melodic sounds in general (Patterson et al., 2002), including pitch contours (Lee et al., 2011). Thus, some of the superior temporal activation may reflect the use of the melodic information to access the musical lexicon. Our results partially replicate those of Peretz et al. (2009) who identified a possible basis of the musical lexicon in the right temporal sulcus. The location of their peak activation in this region is just inferior to the one in our study. Peak coordinates in the study by Peretz et al. (2009) were at MNI x, y, z = 48, −24, −10; the corresponding peak in this study was at MNI x, y, z = 52, −20, 1; Euclidean distance between these two peaks being 12.4 mm.
Other frontal regions of recognition-related activation were in the middle orbitofrontal, superior frontal, and precentral gyri. The orbitofrontal cortex was also identified by Groussard et al. (2010) and Hsieh et al. (2011) for musical memory. This region is a multimodal association area where activity is related to subjective pleasantness (Kringelbach and Radcliffe, 2005) and may therefore, reflect enjoyment of familiar tunes. The superior frontal gyrus, which has been previously associated with musical semantic memory (Platel, 2005; Groussard et al., 2010), may underlie top-down access to knowledge for intentional retrieval (Schott et al., 2005; Binder and Desai, 2011). The precentral gyrus activation may be due to subvocalization, e.g., humming along to familiar music.
Supramarginal activation was found in a study comparing memory for music and odors (Plailly et al., 2007). This region may be involved in maintaining memory for musical pitch (Schaal et al., 2015), and may correspond to recalling the notes of a familiar melody. Subcortical activation in the putamen and brainstem may be associated with motor synchronization to rhythm that is more active when engaging with familiar music (Pereira et al., 2011). It may reflect top-down feedback in the auditory pathway from the cortex to lower nuclei such that familiarity leads to a greater response due to anticipation for the familiar tune. The thalamus may play a similar role for familiar melodies since thalamic activation, having both afferent sensory and efferent motor connections, is linked to task performance, even for very simple tasks, possibly through ongoing changes in motivation and arousal (Schiff et al., 2013).
The main goal of this study was to investigate age differences in the neural basis of melody recognition. Region-of-interest analysis showed that younger adults had greater activation in the left superior temporal gyrus whereas older adults exhibited increased activity in the left superior frontal gyrus and bilateral parietal areas. Because these findings represent new data in the field of aging effects on the neural basis of musical memory, comparisons can only be made to non-musical memory studies where some of the regions involved are distinct from those involved in musical memory. According to the standard consolidation theory, memories become consolidated through stronger cortico-cortico connections over time (Frankland and Bontempi, 2005). The increased activity in the frontal and parietal areas in older adults may result from such consolidation. In a study with famous names, older adults had more activation in the superior frontal and middle temporal regions for “enduring” names (of people who had been for some time and continued to be famous) than non-famous names. Grady (2012) interprets the additional age-related recruitment of frontal areas in general as a compensatory mechanism.
In view of the standard consolidation model of memory, which proposes that memories become less dependent on the medial temporal lobe over time (Frankland and Bontempi, 2005), we expected to find a greater hippocampal region activation in younger adults than in older adults. Younger adults had presumably acquired knowledge of the largely classical and popular tunes more recently than the older adults. A greater activation (marginally significant at p = 0.07) in the right hippocampus of younger adults as opposed to older adults promotes further hypotheses. For example, if the melodies for the adults were from a remote period and unlikely to have been heard more recently, there may be a sharper division between the recent and remote natures of the memories for younger and older adults, respectively. In that case, however, it would be important to ensure a close match, as was done here, in the acoustical and musical properties of the two sets of melodies.
In this study, we have identified the neural basis of melody recognition in a network comprising cortical regions within the superior temporal, inferior and superior frontal, middle orbitofrontal, supramarginal, and precentral gyri. We have shown that older adults engage additional resources in superior frontal and parietal areas whereas younger adults preferentially employ the superior temporal region. Our study makes a reliable contribution to the knowledge of the neural basis of melody recognition based on a large sample size of 40 participants (20 older, and 20 younger, adults). This is also the first study (to the best of our knowledge) to compare brain activity associated with musical semantic memory in younger and older adults.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Grant support was provided by a Discovery Grant from NSERC (RGPIN/333) and a grant from the GRAMMY foundation® to LLC. We thank Dr. Cella Olmstead who was involved in the initial planning for this study, Conor Wild who provided scripts to facilitate data analysis, and Dr. Jordan Poppenk for his extremely valuable feedback.
1. ^These studies of familiar tunes are testing remote (lifetime) memory for music, whereas in memory literature, the terms “familiarity,” “recollection,” and “recognition” most frequently refer to studies where participants study a list of items and are then tested on these items (Yonelinas, 2002), usually in the same session (i.e., a few minutes later). Familiarity in this study-test paradigm refers to knowing whether an item was on the study list. The participant may have lifetime familiarity with ALL the items (e.g., famous faces, or random words), but still not be “familiar” with the item in the sense of having seen it in the study list. This type of memory is likely different than “lifetime familiarity”; the latter is the focus most music recognition studies, including this one.
2. ^All melodies were examined and verified as musically correct by Bernard Bouchard, a composer from the BRAMS Laboratory at the University of Montreal (http://www.brams.org/en/membres/bernard-bouchard/). He also created the stimuli for a similar study on familiar music with younger adults (Peretz et al., 2009).
7. ^The sham MR system has a bore with the same dimensions as the real MRI scanner, as well as a similar table that slides into this opening. The scanning sounds were simulated by playback of the appropriate scan sequence from a CD supplied by the scanner manufacturer. The sound pressure level of the scanning sound simulations was approximately 90 dB, approaching the volume of the MRI scanner.
8. ^The endpoints of this 4-point scale implied that the participant was completely certain, whereas the two points in the middle indicated some uncertainty, about whether they had encountered the tune before. An even number of options allowed us to learn which side of the familiar or unfamiliar question best fit their feelings toward a tune.
10. ^SPM8 from Functional Imaging Laboratory at Wellcome Trust Centre for Neuroimaging, http://www.fil.ion.ucl.ac.uk/spm/, running on MATLAB 2010a from MathWorks, http://www.mathworks.com/products/matlab/.
Ayotte, J., Peretz, I., Rousseau, I., Bard, C., and Bojanowski, M. (2000). Patterns of music agnosia associated with middle cerebral artery infarcts. Brain 123, 1926–1938. doi: 10.1093/brain/123.9.1926
Binder, J. R., Desai, R. H., Graves, W. W., and Conant, L. L. (2009). Where is the semantic system? a critical review and meta-analysis of 120 functional neuroimaging studies. Cereb. Cortex 19, 2767–2796. doi: 10.1093/cercor/bhp055
Brett, M., Anton, J.-L. A., Valabregue, E., and Poline, J.-P. (2002). “Region of interest analysis using an SPM toolbox,” in The 8th International Conference on Functional Mapping of the Human Brain (Sendai).
Douville, K., Woodard, J. L., Seidenberg, M., Miller, S. K., Leveroni, C. L., Nielson, K. A., et al. (2005). Medial temporal lobe activity for recognition of recent and remote famous names: an event-related fMRI study. Neuropsychologia 43, 693–703. doi: 10.1016/j.neuropsychologia.2004.09.005
Folstein, M. F., Folstein, S. E., and McHugh, P. R. (1975). “Mini-mental state”: a practical method for grading the cognitive state of patients for the clinician. J. Psychiatr. Res. 12, 189–198. doi: 10.1016/0022-3956(75)90026-6
Groussard, M., Rauchs, G., Landeau, B., Viader, F., Desgranges, B., Eustache, F., et al. (2010). The neural substrates of musical memory revealed by fMRI and two semantic tasks. Neuroimage 53, 1301–1309. doi: 10.1016/j.neuroimage.2010.07.013
Groussard, M., Viader, F., Landeau, B., Desgranges, B., Eustache, F., and Platel, H. (2009). Neural correlates underlying musical semantic memory. Ann. N. Y. Acad. Sci. 1169, 278–281. doi: 10.1111/j.1749-6632.2009.04784.x
Hébert, S., Peretz, I., and Gagnon, L. (1995). Perceiving the tonal ending of tune excerpts: the roles of pre-existing representation and musical expertise. Can. J. Exp. Psychol. 49, 193–209. doi: 10.1037/1196-19184.108.40.206
Johnson, J. K., Chang, C.-C., Brambati, S. M., Migliaccio, R., Gorno-Tempini, M. L., Miller, B. L., et al. (2011). Music recognition in frontotemporal lobar degeneration and Alzheimer disease. Cogn. Behav. Neurol. 24, 74–84. doi: 10.1097/WNN.0b013e31821de326
Kerer, M., Marksteiner, J., Hinterhuber, H., Mazzola, G., Kemmler, G., Bliem, H. R., et al. (2013). Explicit (semantic) memory for music in patients with mild cognitive impairment and early-stage Alzheimer's disease. Exp. Aging Res. 39, 536–564. doi: 10.1080/0361073X.2013.839298
Lee, Y. S., Janata, P., Frost, C., Hanke, M., and Granger, R. (2011). Investigation of melodic contour processing in the brain using multivariate pattern-based fMRI. Neuroimage 57, 293–300. doi: 10.1016/j.neuroimage.2011.02.006
Mayes, A. R., and Montaldi, D. (2001). Exploring the neural bases of episodic and semantic memory: the role of structural and functional neuroimaging. Neurosci. Biobehav. Rev. 25, 555–573. doi: 10.1016/S0149-7634(01)00034-3
Patterson, R. D., Uppenkamp, S., Johnsrude, I. S., and Griffiths, T. D. (2002). The processing of temporal pitch and melody information in auditory cortex. Neuron 36, 767–776. doi: 10.1016/S0896-6273(02)01060-7
Pereira, C. S., Teixeira, J., Figueiredo, P., Xavier, J., Castro, S. L., and Brattico, E. (2011). Music and emotions in the brain: familiarity matters. PLoS ONE 6:e27241. doi: 10.1371/journal.pone.0027241
Peretz, I., Gosselin, N., Belin, P., Zatorre, R. J., Plailly, J., and Tillmann, B. (2009). Music lexical networks: the cortical organization of music recognition. Ann. N. Y. Acad. Sci. 1169, 256–265. doi: 10.1111/j.1749-6632.2009.04557.x
Peretz, I., Kolinsky, R., Tramo, M., Lebrecque, R., Hublet, C., Demeurisse, G., et al. (1994). Functional dissociations following bilateral lesions of auditory cortex. Brain 117, 1283–1301. doi: 10.1093/brain/117.6.1283
Platel, H., Baron, J.-C., Desgranges, B., Bernard, F., and Eustache, F. (2003). Semantic and episodic memory of music are subserved by distinct neural networks. Neuroimage 20, 244–256. doi: 10.1016/S1053-8119(03)00287-8
Satoh, M., Takeda, K., Nagata, K., Shimosegawa, E., and Kuzuhara, S. (2006). Positron-emission tomography of brain regions activated by recognition of familiar music. Am. J. Neuroradiol. 27, 1101–1106.
Schaal, N., Williamson, H., Maria, S., Neil, L., Neuroscience, C., and Banissy, M. J. (2015). A causal involvement of the left supramarginal gyrus during the retention of musical pitches. Cortex 64, 310–317. doi: 10.1016/j.cortex.2014.11.011
Schiff, N. D., Shah, S. A., Hudson, A. E., Nauvel, T., Kalik, S. F., and Purpura, K. P. (2013). Gating of attentional effort through the central thalamus. J. Neurophysiol. 109, 1152–1163. doi: 10.1152/jn.00317.2011
Schott, B. H., Henson, R. N., Richardson-Klavehn, A., Becker, C., Thoma, V., Heinze, H.-J., et al. (2005). Redefining implicit and explicit memory: the functional neuroanatomy of priming, remembering, and control of retrieval. Proc. Natl. Acad. Sci. U.S.A. 102, 1257–1262. doi: 10.1073/pnas.0409070102
Smith, C. N., and Squire, L. R. (2009). Medial temporal lobe activity during retrieval of semantic memory is related to the age of the memory. J. Neurosci. 29, 930–938. doi: 10.1523/JNEUROSCI.4545-08.2009
Tzourio-Mazoyer, N., Landeau, B., Papathanassiou, D., Crivello, F., Etard, O., Delcroix, N., et al. (2002). Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject brain. Neuroimage 15, 273–289. doi: 10.1006/nimg.2001.0978
Vanstone, A. D., Sikka, R., Tangness, L., Sham, R., Garcia, A., and Cuddy, L. L. (2012). Episodic and semantic memory for melodies in Alzheimer's disease. Music Percept. 29, 501–507. doi: 10.1525/mp.2012.29.5.501
Weinstein, J., Koenig, P., Gunawardena, D., McMillan, C., Bonner, M., and Grossman, M. (2011). Preserved musical semantic memory in semantic dementia. Arch. Neurol. 68, 248–250. doi: 10.1001/archneurol.2010.364
Worsley, K. J., Marrett, S., Neelin, P., Vandal, A. C., Friston, K. J., and Evans, A. C. (1996). A unified statistical approach for determining significant signals in images of cerebral activation. Hum. Brain Mapp. 4, 58–73.
Yee, E., Chrysikou, E. G., and Thompson-Schill, S. L. (2013). “The cognitive neuroscience of semantic memory,” in Oxford Handbook of Cognitive Neuroscience, Vol. 1 Core Topics, eds K. Ochsner and S. Kosslyn (New York, NY: Oxford University Press), 353–374.
Keywords: recognition, familiarity, fMRI, music, aging, semantic memory, melodies
Citation: Sikka R, Cuddy LL, Johnsrude IS and Vanstone AD (2015) An fMRI comparison of neural activity associated with recognition of familiar melodies in younger and older adults. Front. Neurosci. 9:356. doi: 10.3389/fnins.2015.00356
Received: 04 June 2015; Accepted: 17 September 2015;
Published: 06 October 2015.
Edited by:Einat Liebenthal, Brigham and Women's Hospital/Harvard Medical School, USA
Reviewed by:Petr Janata, University of California, Davis, USA
Howard Charles Nusbaum, The University of Chicago, USA
Copyright © 2015 Sikka, Cuddy, Johnsrude and Vanstone. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Ritu Sikka, Music Cognition Lab, Queen's University, 62 Arch Street, Humphrey Hall Room 232, Kingston, ON K7L 3N6, Canada, firstname.lastname@example.org