- 1Laboratory of Behavioural Neurodynamics, Saint-Petersburg State University, Saint-Petersburg, Russia
- 2Affective Psychophysiology Laboratory, Institute of Health Psychology, National Research University Higher School of Economics (HSE), Saint-Petersburg, Russia
Although the difference between consonance and dissonance has raised interest for decades in various fields of science, isolated chords are still underutilised as standardised affective stimuli in neuroscience. In the present study, we assessed whether consonant, dissonant, and neutral sounds evoked different subjective and neurophysiological responses associated with the emotional experience. For the first time, we conducted a comprehensive study of piano timbre chord perception, combining behavioural assessments, event-related potentials (ERPs), frequency-domain electroencephalography (EEG) analysis of amplitude, neurodynamic (longrange temporal correlations (LRTC)) factors, and source analysis. In an experiment, 30 participants rated the valence of sounds (consonance, dissonance, neutral) whilst simultaneously undergoing EEG recordings. At the behavioural level, a stable valence gradient was revealed: consonance was perceived as pleasant, dissonance as unpleasant, and the neutral group of stimuli as neutral. Analysis of ERPs revealed differences in response across three time windows (90–110 ms, 190–210 ms, 290–310 ms) and activation of frontotemporal and temporoparietal areas during the processing of dissonant chords. At the level of frequency and neurodynamic indices (gamma and beta bands amplitudes and LRTC), we demonstrated that classification accuracy depends on the interaction between chord type and the EEG’s amplitude- and scale-invariant characteristics for dissonant and neutral stimuli. These results provide evidence that isolated chords evoke differentiated emotional and cognitive responses, highlighting their potential utility as affective stimuli in future studies.
1 Introduction
The dichotomy between consonance and dissonance, originating in Pythagorean theory and subsequently reinterpreted through the Platonic prism of cosmic harmony (Lippman, 1994; Plato, 2019), remains a relevant theoretical origin underlying works ranging from music theory to cognitive neuroscience. Within this trajectory, musical intervals defined by simple frequency ratios (e.g., octaves, fifths) have been considered as pleasant due to their stability. Dissonant intervals, on the other hand, were characterised by more complex frequency relationships and were considered acoustically unstable, thus affecting the listener unpleasantly (Krumhansl and Jusczyk, 1990; Bidelman and Krishnan, 2009).
This differential perception of pleasant and unpleasant intervals is a specific manifestation of a broader neurophysiological mechanism for evaluating auditory stimuli. Indeed, emotional processing of distant stimuli, such as auditory or visual stimuli, is fundamental to the human experience and is associated with behavioural and physiological level reactions (Bradley and Lang, 2000; Keil et al., 2002). Negative auditory stimuli, such as screeching or screaming, typically evoke autonomic body responses attributed to the activation of the sympathetic nervous system (Kumar et al., 2012; Schweiger Gallo et al., 2017). Concrete sounds, however, are typically associated with specific, dangerous objects or circumstances related to traumatic episodes in a person’s life. Music, being more abstract than environmental sounds and rooted in culture, offers a different type of emotional impact that engages distributed networks, including limbic and default mode systems (Koelsch, 2010; Brattico et al., 2011; Schirmer, 2018).
Notably, affective research paradigms have long been dominated by visual stimuli, particularly static images, such as emotional faces or scenes (Lang and Bradley, 2010; Carretié, 2014; Dawel et al., 2022), and dynamic videos (Samson et al., 2016; Maffei and Angrilli, 2019). Whilst static images reliably evoke isolated unpleasant reactions (e.g., fear or disgust), videos induce higher arousal and more varied emotional responses (Horvat et al., 2015). However, in terms of temporal accuracy, ecological validity, and cultural specificity, both classes of visual stimuli have meaningful limitations (Boğa et al., 2023).
In contrast, auditory stimuli provide enhanced opportunities for cross-modal integration in experimental settings. Thus, amongst emotional primitives, consonance and dissonance are typically associated with pleasantness and unpleasantness, respectively (Fritz et al., 2009; Virtala and Tervaniemi, 2017; Harding et al., 2025). Auditory stimuli also elicit more stable and extensive autonomic responses when combined with visual stimuli (Song et al., 2021). In addition, validated auditory affective paradigms such as the International Affective Digitised Sounds (IADS) database (Bradley and Lang, 2000; Yang et al., 2018) and standardised music-emotion protocols (Juslin and Västfjäll, 2008; Koelsch, 2014) provide robust frameworks for studying affective auditory processing. Although less frequent than visual paradigms (Gerdes et al., 2014), these approaches highlight the critical role of auditory stimuli in emotion research.
Importantly, affective responses to consonance and dissonance agree with the findings of prosody research, in which the emotional characteristics of a speech stimulus are associated with variations in pitch and timbre (Schirmer and Kotz, 2006; Paulmann and Kotz, 2008; Liu et al., 2018). Both processes are based on mechanisms that rapidly evaluate acoustic characteristics, involving overlapping neural networks in the superior temporal and frontal cortices of the brain (Proverbio et al., 2020). In addition to these parallels, theoretical frameworks such as predictive coding characterise both music and speech as hierarchical systems where deviations from expected patterns—whether harmonic or prosodic—trigger prediction error signals (Friston, 2005; Koelsch et al., 2019). The shared syntactic integration resource hypothesis (Patel, 2003) suggests that there is a partially common pool of resources primarily linked to Broca’s area. Neuroimaging studies indicate that emotional prosody and dissonant harmonies activate overlapping frontal-temporal networks (Schirmer and Kotz, 2006; Koelsch, 2014; Proverbio et al., 2020). Furthermore, both limbic and auditory association areas are engaged by emotional vocalisations and harmonic expectancy (Frühholz et al., 2014). This similarity between emotional evaluation mechanisms for chords and prosodic speech signals highlights their potential as standardised stimuli for shared mechanisms of emotion across music and language.
Despite these advantages, the implementation of isolated musical chords within the domain of affective neuroscience remains limited. Chords represent a compact but powerful type of stimuli: parametrically driven, affectively salient, and ecologically relevant (in cases where a stimulus is required that is not associated with a specific object or situation). However, it remains debated whether these emotional labels correspond to genuine affective states or are the result of cognitive categorisation based on acoustic characteristics, such as frequency ratio or roughness (Gabrielsson, 2001; Eerola and Vuoskoski, 2011), and cultural experience (Parncutt and Hair, 2018). Recent meta-analyses of consonant and dissonant sounds suggest that simultaneous consonance is a complex phenomenon that largely stems from three key factors: interference, periodicity/harmonicity, and cultural familiarity. Additionally, the perceptual salience of dissonance has been also related to roughness as a marker of potential danger (Harrison and Pearce, 2020; see, also, Di Stefano et al., 2022).
Empirical studies have shown that harmonic structure modulates the auditory components of the event-related potentials (ERP): consonant stimuli often evoke an increase in N1 amplitude with latency around 100 ms, whereas dissonant stimuli are associated with an enhanced P2 response with latency around 200 ms, suggesting higher auditory coding demands (Regnault et al., 2001; Bidelman and Grall, 2014). Comparative studies have shown that both musicians and non-musicians exhibit early deviance-related potentials to unexpected chords such as MMN or its music-syntactic analogue ERAN (Koelsch et al., 2000; Vuust et al., 2012; Quiroga-Martinez et al., 2022). Taken together, N1, P2, and MMN represent well-established early ERP markers of auditory-affective processing, indexing rapid perceptual encoding, categorisation, and deviance detection that provide the neural basis for subsequent evaluative components. Notably, several studies have shown that these early components can display a “negativity bias,” with unpleasant or threatening stimuli eliciting stronger responses than arousal-matched positive ones (Carretié et al., 2004; Delplanque et al., 2006; Olofsson et al., 2008). Electrocorticographic (ECoG) recordings emphasise the sensitivity of the right superior temporal region to dissonance in the high gamma band between 75 and 200 ms post-stimulus onset (Foo et al., 2016).
The electrophysiological correlates of affective evaluation in response to musical stimuli are less well characterised than those underlying perceptual or syntactic processing. Most ERP studies have focused on amplitude rather than latency; arousal effects are typically observed at longer latencies, whereas findings on valence effects remain less consistent (Olofsson et al., 2008). Two main ERP components modulated by stimulus arousal are the early posterior negativity (EPN), with a peak on 200–300 ms after stimulus onset, which probably indicates the selection of emotionally arousing stimuli for further processing; this component has also been discussed in relation to predictive coding frameworks (Friston, 2005). The second one is the late positive potential (LPP), a slow-wave component associated with consolidation of emotional memory (Brown et al., 2012). Previously, LPP has been reported to be larger for negatively valenced stimuli (Ito et al., 1998).
Recent EEG and MEG studies on music perception demonstrate that chord sequences elicit specific ERP components related to both cognitive and affective evaluations. Thus, chords evaluated as ‘incorrect’ often elicit an early right anterior negativity (ERAN) with latency 150–200 ms after stimulus onset or later (≈250–280 ms) in some unpredictable-sequence paradigms, corresponding to pre-attentive neural responses to harmonic violations in Western tonal music, rather than subjective listener judgements (Ellison et al., 2015; Ishida and Nittono, 2022). Additionally, negative evaluations (‘dislike’ or ‘incorrect’) of musical chords are associated with centro-parietal P3b/P600 positivity in ERP signal around 400–600 ms, indicative of conscious cognitive evaluation. However, this remains distinct in the context of music perception. Finally, an LPP around 1,200 ms was observed, reflecting higher-order cognitive processes related to the conscious decision-making regarding musical correctness (in music harmony terms) or preference, rather than merely emotional arousal (Brattico et al., 2010; Zhang et al., 2022).
Additionally, in conjunction with ERP, another essential characteristic of the EEG that determines a reaction to external signals is the average brain activity within a particular frequency range of spectral activity oscillations, as well as its dynamics (Li et al., 2018). Since such activity of a specific person is relatively stable compared to the ERP, there is even a tendency to create a personal EEG profile (Sysoeva et al., 2023). Such a personal profile may be associated, amongst other things, with the individual’s characteristics and their reaction to specific emotional stimuli. There is evidence that different frequency bands of the brain are associated with particular brain functions. Although there is considerable discussion in the literature about this issue, is it possible to link a specific oscillation frequency to the functional activity of the brain? For example, previous studies have linked theta oscillations to affective processing and motivational salience (Miltner et al., 1999), alpha rhythms to inhibition and attentional control in emotion regulation, beta to the regulation of emotional responses (Balconi and Lucchiari, 2008), and gamma activity to emotional arousal (Li and Lu, 2009). In more detail, oscillatory brain activity in various EEG frequency bands provides meaningful insight into the temporal and functional dynamics of emotional information processing. According to recent reviews, slow delta (0.5–4 Hz) and theta (4–7 Hz) rhythms are increased when a stimulus has high motivational significance and are associated with emotion regulation. Frontal theta oscillations increase during cognitive control in conditions of emotional conflict and successful re-evaluation of the situation (Knyazev, 2007; Ertl et al., 2013; Güntekin and Başar, 2014). Alpha activity (8–13 Hz) reflects the mechanisms of inhibition and attention distribution: a decrease in alpha power in the frontal lobes corresponds to high cortical activation during an emotional task, and frontal asymmetry of the alpha rhythm (FAA; ln R – ln L) correlates with approach and avoidance motivation, although the reliability of this index varies between studies (Davidson, 1998; Allen et al., 2004; van der Vinne et al., 2017). Beta rhythms (13–30 Hz), associated with arousal and motor readiness, show an increase in both anxiety states and increased emotional control (Miskovic and Schmidt, 2010; Schubring and Schupp, 2021). It is important to note that the functional significance of beta activity in emotional contexts remains a subject of debate. The literature reports seemingly contradictory findings, where beta power has been associated with processes ranging from anxiety and emotional regulation (Miskovic and Schmidt, 2010) to sensory integration and cognitive evaluation (Schubring and Schupp, 2021). This variability suggests that beta oscillations are not a univariate marker of a single process such as anxiety but may reflect a broader mechanism of cortical control or active inhibition during salient perceptual events. Gamma activity (>30 Hz) increases during highly arousing emotional stimuli and reflects integrative evaluative processing in occipito-temporal and limbic regions (Keil et al., 2002; Balconi and Lucchiari, 2008; Strube et al., 2021). Recent evidence suggests that enhancement of gamma potentials during the presentation of affective stimuli may be associated with improved subsequent memory, particularly for highly arousing negative scenes (Theódórsdóttir and Höller, 2023). However, this finding has thus far been reported primarily in the context of visual paradigms, and further research is needed to establish whether it generalises to auditory affective stimuli. Additionally, Güntekin and Başar (2014) provided a narrative overview of oscillatory responses to emotional visual stimuli, highlighting frequency-specific modulations but without applying a systematic methodology. Moreover, over the past few decades, new methods for assessing the neurodynamics of various rhythms have been developed, such as a method for analysing long-term correlations in the EEG signal (Sangiuliano Intra et al., 2018). This method has a physical basis, since it reflects the excitation/inhibition ratio, as well as the efficiency of information processing in the nervous system (Nikulin and Brismar, 2004, 2005). The relationship between long-range temporal correlations (LRTC) and emotions can be specific to certain EEG frequency bands (e.g., theta, alpha, beta), with different bands potentially reflecting various aspects of emotional processing (Bornas et al., 2015). In this study, we also assessed the emotional perception of different chords based on the frequency analysis of the subject’s EEG spectrum and the neurodynamics (LRTC) of the main rhythms.
This study aims to expand on existing work by employing a comprehensive methodology, with a particular focus on two aspects: first, the underutilization of auditory consonant stimuli, particularly chords, as affective stimuli; and second, the relatively modest integration of such stimuli into multimodal affective neuroscience research. We hypothesise that isolated musical chords will elicit reliable behavioural and electrophysiological markers of emotion—specifically, differentiated self-reported valence/arousal ratings and modulations of emotion-sensitive ERP components such as early components (N1, N2, P3). We also employed an analysis that combined behavioural categorisation with EEG recordings for source localisation (sLORETA) to investigate the early and late stages of affective auditory processing of consonant and dissonant chords. Based on previous data (Pascual-Marqui, 2002), we hypothesised that emotional chords would evoke enhanced P2 and P3a components compared to neutral sounds, reflecting their higher emotional salience and attention capture. We also assume that a person’s response to certain stimuli may be associated with different EEG frequency ranges, both in their amplitude and neurodynamic components (LRTC). By considering consonance and dissonance within the broader framework of affective neuroscience and grounding our predictions in psychophysiological and cultural theory, this study proposes musical chords as compact, emotionally relevant, and experimentally accessible stimuli for studying the neurocognitive architecture of affect.
2 Materials and methods
2.1 Participants
Thirty healthy non-musicians (mean age = 25.6 years, SD = 8, range = 18–45, 21 females) participated in the study. All had no history of hearing problems, speech disorders, neurological or psychiatric conditions, and no use of medications affecting brain function, as confirmed by a brief screening questionnaire. Participants’ musical experience was assessed in a short, structured oral interview. They were asked to report (i) whether they had ever received formal musical training, (ii) the instrument(s) played, (iii) the total number of years of instruction, (iv) whether they were currently practising or performing, and (v) the approximate frequency of musical activity in the past 5 years. Participants were not considered as musicians if they had less than 2 years of musical education and had not engaged in regular rehearsals (defined as more than 1 h per week) during the previous 5 years. These thresholds are broadly consistent with prior studies (e.g., Koelsch et al., 2000; Bidelman and Krishnan, 2009). All participants gave their written consent to participate in the experiment. The Ethics Committee of St. Petersburg State University approved the study.
2.2 Experimental task and procedure
In the present study, auditory stimuli were constructed using piano tones from the C₃–B₃ range (130.8–246.9 Hz, traditionally referred to as the ‘small octave’) and from the C₄–B₄ range (261.6–493.9 Hz, traditionally referred to as the ‘one-line octave’). All chords were constructed in accordance with the method described by Steinbeis and Koelsch (2011) and were used as separate individual chord sounds. The consonant chords consisted of seven major triads in root position, presented in the configuration I–V–I (C–E–G–C, D–F#–A–D, E–G#–B–E, F–A–C–F, G–B–D–G, A–C#–E–A, B–D#–F#–B) and seven six-four inversions of these chords in the V–I–V position (e.g., G–C–E–G for C major, A–D–F#–A for D major).
Dissonant chords were constructed by superimposing fixed interval patterns on the root notes of the natural diatonic scale (C, D, E, F, G, A, B). Two types of dissonant structures were used. The first consisted of the root, augmented fourth (+6 semitones), perfect fourth (+5 semitones), and minor second (+1 semitone), for example: C–F#–B–C, D–G#–C#–D, E–A#–D#–E. The second consisted of the root, minor second (+1 semitone), perfect fourth (+5 semitones), and augmented fourth (+6 semitones), for example, C–C#–F#–C, D–D#–G#–D, E–F–A–E. These interval structures were designed to maximise sensory dissonance through the proximity of sounds and non-harmonic intervals. All stimuli consisted of isolated chords defined by their intervallic content, with no harmonic progression intended or implied.
All sounds were generated using GarageBand software (Apple Inc., Cupertino, CA, United States). All chord stimuli had a duration of 800 ms and were normalised in amplitude using GarageBand software. They were synthesised with the same timbre (Grand Piano preset). The neutral auditory control consisted of single metronome clicks. In affective ERP paradigms, neutral stimuli are typically operationalised as low-complexity baseline conditions (Bradley and Lang, 2000; Olofsson et al., 2008) that minimise emotional salience. In line with this principle, we chose metronome clicks, which are acoustically simple, culturally familiar, and free of harmonic content. These clicks were synthesised using a sampler and a pre-recorded metronome sound. Each digitally metronome sound was modulated in pitch to correspond to a tone within the C₃–B₄ range (the small and one-line octaves). Auditory stimuli were presented via semi-open headphones (AKG K-240 mk2, Harman International Industries) at an individually adjusted, subjectively comfortable volume level.
Throughout the experimental task (Figure 1), a fixation cross remained continuously displayed at the centre of the screen to maintain participants’ visual attention and minimise eye movements. Auditory stimuli were presented separately with 200–300 ms interval between stimuli via high-quality stereo headphones, ensuring balanced binaural delivery. Each auditory stimulus was presented five times across the session in a randomised order, whilst the visual fixation cross was presented on the screen, the total duration of the experimental session did not exceed 20 min. Following each stimulus presentation, participants were instructed to categorise their subjective impression of the sound as pleasant, unpleasant, or neutral. Responses were provided using the index finger of the right hand on a response pad (RB-740, Cedrus Inc.), with each key corresponding to one of the three response options (neutral, pleasant or unpleasant). Throughout the task, continuous recordings of brain activity with EEG, reaction times, and behavioural responses were collected for subsequent analysis. At the onset of each auditory stimulus, a time-locked event marker was sent to the EEG recording system, allowing for the precise synchronisation of neural and behavioural data.
2.3 Recording
EEG data were recorded using a BrainVision Recorder (BrainProducts GmbH, Gilching, Germany) with a 64-channel LiveAmp (BrainProducts GmbH, Gilching, Germany) system and an R-Netcap (BrainProducts GmbH, Gilching, Germany) system. The FCz electrode was used as the reference. Electrode placement followed the extended 10–20 system (M1 montage; EasyCap, Herrsching, Germany), with all impedances kept below 25 kΩ. Signals were sampled at 500 Hz without online filtering. During the experimental procedure, participants were sitting in a comfortable, adjustable chair inside a sound-attenuated, temperature-controlled, and shielded room.
2.4 EEG analysis
2.4.1 EEG preprocessing
EEG data were preprocessed using a custom MATLAB-based pipeline (R2022b, MathWorks, Natick, MA, United States), in combination with the Berlin Brain-Computer Interface (BBCI) (https://github.com/bbci) toolbox. EEG recordings were first downsampled to 250 Hz to optimise preprocessing (e.g., ICA) and reduce computational load, without loss of information in the analysed frequency range (≤45 Hz). A zero-phase 2nd-order Butterworth band-stop filter (45–55 Hz) was applied as a notch filter, followed by a band-pass filter (1–45 Hz). Data were subsequently re-referenced to the common average reference.
Noisy channel detection was performed through spectral power analysis and visual inspection, with channels exhibiting excessive noise (absolute amplitude greater than 100 μV). Independent Component Analysis (ICA) was conducted using FastICA (Hyvarinen, 1999). Components associated with ocular artefacts were identified by their characteristic frontal topography and temporal correlation with blink events and subsequently removed from the data.
Preprocessed EEG data were segmented into epochs time-locked to stimulus onset (−100 to 800 ms) and baseline-corrected using the −100 to 0 ms pre-stimulus interval.
2.4.2 Source analysis
Finally, given the observed differences in ERP amplitudes across chord types, we computed the neural generators underlying this ERP activity using a distributed source reconstruction approach. Specifically, an sLORETA solution (Pascual-Marqui, 2002) was calculated on a realistic head model based on the Montreal Neurological Institute (MNI) template, employing a boundary-element method to account for the conductivity properties of head tissues. Source reconstruction was constrained to cortical grey matter and applied to grand-averaged data to enhance signal-to-noise ratio. The analysis was performed using the sLORETA/eLORETA software package on the mean ERP of dissonant and neutral conditions within the peak time intervals identified in the signal-space ERP analysis. sLORETA has a fundamental limitation in its low spatial resolution, but with the number of electrodes 64 is enough to estimate source localisation.
2.4.3 Amplitude analysis and long-range temporal correlations
To assess the amplitude and neurodynamics of the main EEG rhythms, the entire EEG recording was used throughout the experiment. The amplitude of the EEG signal in each frequency range was estimated using the following algorithm: the signal was filtered in certain frequency ranges (alpha: 8–12 Hz, beta: 20–25 Hz, gamma: 35–40 Hz) using a two-way Butterworth bandpass filter. For the resulting signal, an envelope curve was constructed based on the Gilbert transform. To assess the amplitude of each rhythm, the average value of the envelope curve was found. To evaluate the neurodynamic characteristics, detrended fluctuation analysis (DFA) was employed, which reflects the LRTC with the window length varying from 5 to 50 s distributed equidistantly on a logarithmic scale (the algorithm for this analysis is described in Nikulin and Brismar, 2004).
2.4.4 Statistical analysis
We applied Fisher’s exact test (Fisher, 1992) to examine the association between stimulus type and response category. To compare reaction times (RT) across the six groups formed by stimulus type (consonant, dissonant, neutral) and response congruency (congruent vs. incongruent), we then used a Kruskal–Wallis rank-sum test (Kruskal and Wallis, 1952). In all pos-hoc analysis Bonferroni correction was used.
All EEG/ERP analyses were conducted using custom scripts developed in the MATLAB 12.0 environment (MathWorks, Natick, MA), in combination with the BBCI toolbox. To address the issue of multiple comparisons arising from tests conducted across numerous channels, statistical significance for both the Wilcoxon signed-rank tests (Wilcoxon, 1945) and Spearman correlations (Spearman, 1904) was assessed across subjects using cluster-based permutation methods (Maris and Oostenveld, 2007). The permutation statistics is a method for multiple comparison analysis in electrode space. Eight hundred permutations were conducted for each contrast, with a two-tailed significance threshold set at p < 0.05. Cluster-level statistics were corrected for multiple comparisons using the family-wise error rate (FWER), allowing for robust inference on the spatiotemporal patterns of EEG activity.
Generalised linear mixed-effects models (GLMMs) with a binomial logit linkage (Hensher and Greene, 2003) were used to assess how oscillatory neural activity modulates affective categorisation. GLMMs are a flexible statistical framework for analysing binary outcomes. They allow the inclusion of both continuous and categorical predictors, whilst considering random effects such as variability at the participant and stimulus levels. The models included fixed effects for stimulus type (consonant, neutral, dissonant), oscillation dynamics in relevant frequency bands (DFA and frequencies), and their interactions. To account for repeated measures and individual variability, the model specified random intercepts and slopes for stimulus type across participants, and random intercepts across items (i.e., unique chord stimuli).
This modelling approach allowed us to test whether the probability of categorising stimuli as a priori groups (consonant as pleasant, dissonant as unpleasant, neutral as neutral) changes as a function of stimulus type and neural dynamics during the task. We used likelihood ratio tests to compare full models with reduced (null) models and calculated odds ratios with 95% confidence intervals from fixed effects to simplify interpretation.
In our interpretation, we focused on interaction effects directly from the retrieved models. This decision was based on current best practises in statistical modelling of neurocognitive data, emphasising the interpretability and simplicity of theoretically based models (Barr et al., 2013; Schad et al., 2020).
We did not apply a global correction across all analyses (ERP, oscillatory, LRTC, and behavioural); the results should therefore be interpreted accordingly.
3 Results
3.1 Behavioural results
Behavioural results showed significant differences in response distributions across consonant, dissonant, and neutral auditory stimuli. Current analysis was based on predefined expected responses: “pleasant” for consonant chords, “unpleasant” for dissonant chords, and “neutral” for metronome sounds.
The analysis demonstrated a statistically significant association between stimulus type and response classification (p < 0.001), indicating that the distribution of correct and incorrect responses varied systematically across stimulus categories (Figure 2). These results suggest that participants statistically differentiate the emotional valence of consonant, dissonant, and neutral auditory stimuli with statistical significance.
 
  Figure 2. Proportion of “pleasant,” “neutral,” and “unpleasant” responses in categorisation task for dissonant, consonant, and neutral chords.
A Kruskal–Wallis rank sum test revealed a significant difference in RT between groups, χ2(5, N = 4,702) = 261.99, p < 0.001. Post-hoc pairwise comparisons using Wilcoxon rank-sum tests with Bonferoni correction showed significant differences (adjusted p < 0.05) between most conditions, except for comparisons between dissonant and consonant in each congruent and incongruent condition, which were not statistically significant. Mean reaction times were as follows: dissonant congruent—522 ms, consonant congruent—464 ms, neutral congruent—747 ms, dissonant incongruent—1,123 ms, consonant incongruent—1,158 ms, neutral incongruent—688 ms.
3.2 EEG results
A global field power transformation (GFP) was performed across all stimuli, conditions, and volunteers to objectively quantify the overall response pattern and identify time windows for further analysis. The resulting GFP curve showed three most prominent peaks corresponding to classical ERP components with maxima at ~100, 200, and 300 ms. Time windows of 90–110 ms, 190–210 ms, and 290–310 ms post-stimulus onset were identified as regions of interest based on the timing of overall neural activity (Figures 3A,B).
 
  Figure 3. EEG results: (a) Global field power (GFP) of the ERP responses across all scalp electrodes for consonant, dissonant and neutral audio stimuli with three time-windows of interest 90–110, 190–210, and 290–310 ms after stimulus onset; (b) ERP responses in FCz electrode for consonant, dissonant and neutral audio stimuli; (c) Results of permutation analysis is shown clusters of electrodes with statistically significant differences between ERP responses in different stimulus types in considered time windows. Red crosses indicate channels belonging to a significant cluster (p < 0.05). The coloured scale (in μV) on the figure represents the gradient of the difference on the topograms.
Permutation analysis (Figure 3C) was used to identify significant differences in brain activation between stimulus groups: consonant versus neutral chords, dissonant versus neutral chords, and consonant versus dissonant chords. Difference curves were calculated for each pair of conditions, followed by permutation analysis to determine statistically significant activation differences in the selected time windows.
In the earliest time window (90–110 ms), comparisons revealed distinct scalp activation patterns for different stimulus conditions. The contrast between consonant and neutral chords revealed significant differences, predominantly in the posterior-central and right anterior-central scalp regions. Comparing dissonant chords with neutral chords identified localised significant differences in posterior occipital scalp areas. Additionally, consonant versus dissonant chord comparisons demonstrated moderate but focal significant differences in the right temporal and left central scalp regions. In the time window 190–210 ms, dissonant chords compared to neutral chords elicited widespread significant activation differences, predominantly observed over bilateral posterotemporal and right frontocentral scalp regions. In the time window 290–310 ms, comparisons between dissonant and neutral chords revealed extensive and significant activation differences distributed broadly across bilateral frontal, temporal, and occipital scalp areas.
To further localise and interpret the neural generators of these ERPs, sLORETA analysis was conducted (Figure 4). Source localisations were assessed separately for dissonant and neutral stimulus types across the two latest time windows. As the analysis was exploratory and the spatial precision of sLORETA is limited, no quantitative peak values or coordinates are reported.
 
  Figure 4. Cortical source activation contrasts between dissonant and neutral auditory stimuli across two analysed time windows (190–210 ms and 290–310 ms).
Brain activation observed at the first time window (90–110 ms) was excluded from the detailed analysis due to its relatively small magnitude differences and limited relevance to the emotional processing objectives of this study. At the earlier interval (190–210 ms), dissonant stimuli elicited enhanced cortical activation predominantly in the right frontotemporal and bilateral posterior temporoparietal areas compared to neutral stimuli, indicating increased cognitive and affective processing specifically associated with dissonance perception. During the later interval (290–310 ms), the response to dissonant stimuli further intensified, resulting in widespread bilateral frontal and temporal activation suggestive of heightened attentional engagement and emotional salience. Notably, the pattern of source lateralisation evolved from predominantly right-hemispheric during the earlier stage to symmetrical bilateral involvement in the later processing stage.
3.3 Mixed-effects logistic regression: oscillatory predictors of accuracy
To assess the relationship between trial-level categorisation accuracy and oscillatory brain activity, we fitted a series of mixed-effects logistic regression models. Each model included fixed effects for stimulus type (consonant, dissonant, neutral) and a continuous neural predictor (either group-specific amplitude or LRTs based on DFA), as well as their interaction. Random intercepts were set for stimuli, and random intercepts and uncorrelated slopes for stimulus type were included for participants. The significant and marginal effects found in the mixed-effects logistic models are presented in Table 1. Full details of the additional model analyses are provided in Supplementary materials S4–S7.
 
  Table 1. Significant and marginal effects from mixed-effects logistic models predicting classification accuracy.
3.3.1 Amplitude-based models
In the alpha band (8–12 Hz), neither the main effect of amplitude nor the interaction with stimulus type reached statistical significance (all p < 0.05).
In the beta band (20–25 Hz), a significant interaction was found between beta amplitude and dissonant stimuli (β = −2.08, SE = 0.94, z = −2.22, p = 0.027, OR = 0.12 [0.02, 0.78]), indicating that increased beta amplitude was associated with reduced classification accuracy for dissonant chords.
In the gamma band (35–40 Hz), dissonant stimuli were more likely to be correctly classified when gamma amplitude was low (β = 2.13, SE = 0.84, z = 2.52, p = 0.012, OR = 8.41 [1.6, 44]), but this effect reversed at higher gamma amplitudes (β = −4.37, SE = 1.49, z = −2.93, p = 0.003, OR = 0.01 [0.00068, 0.23]).
3.3.2 DFA-based models
In the alpha band DFA model, no main effects or interactions were statistically significant (all ps ≥ 0.12).
For beta DFA (20–25 Hz), classification accuracy was higher for neutral stimuli (p = 0.074), although this advantage diminished at higher beta DFA values (β = −8.10, SE = 4.13, z = −1.96, p = 0.050, OR < 0.001 [9.1e-08, 1]).
In the gamma DFA logistic regression model (35–40 Hz), neutral chords were more accurately classified at low DFA values (β = 4.88, SE = 2.03, z = 2.40, p = 0.016, OR = 131.63 [2.4, 7100]) [an extreme odds ratio followed by a very wide confidence interval reflects an imbalance in the number of correct and incorrect responses in this type of presented material and we interpret this outcome cautiously, focusing on the direction of the association (OR > 1)]; however, this effect was reversed with increasing gamma DFA (β = −9.74, SE = 3.07, z = −3.17, p = 0.002, OR < 0.001 [1.4e-07, 0.024]).
4 Discussion
The current study investigated whether isolated musical chords in piano timbre, differing in harmonic structure (consonant, dissonant, or neutral), or neutral metronome clicks evoke differentiated affective responses in non-musicians at both behavioural and neurophysiological levels, including ERP (with source analysis), frequency, and neurodynamics analysis.
Our results showed differences in behavioural valence ratings: consonant chords were rated as more pleasant. In contrast, dissonant chords were rated as more unpleasant, whereas neutral metronome clicks were more often rated as neutral; thus supporting a reliable distinction between consonant and dissonant chords and metronome clicks in subjective valence ratings. Consonant chords were rated as more pleasant than dissonant chords, which, in turn, were rated as more unpleasant. These results replicate earlier findings that consonance and dissonance are categorised as pleasant or unpleasant, respectively (Gabrielsson, 2001; Bidelman and Krishnan, 2009; McDermott et al., 2010). Furthermore, in our study, consonant and dissonant chords differed in their ratings from neutral auditory stimuli, indicating a clear differentiation of the selected auditory stimuli. It is noteworthy that the average ratings of neutral stimuli indicate that listeners did not rely on binary judgements but differentiated stimuli according to valence gradation. The precise pattern of subjective differentiation obtained even amongst non-musician listeners supports the idea that isolated harmonic structures alone are sufficient for evaluative affective judgements. However, it should be noted that previous studies have demonstrated that listeners can make valence judgements based on acoustic features such as roughness or interval size without necessarily experiencing an affective state (Gabrielsson, 2001; Juslin and Laukka, 2004).
To investigate the experience of emotions, an EEG analysis was conducted. At the electrophysiological level, different ERP components were modulated by chord type: all three sound types differed at the level of primary auditory processing, and at later stages of processing, dissonant chords evoked greater P2 and P3a amplitudes compared to neutral stimuli, indicating increased attentional and affective salience. Consider these results in more detail during the P1/N1 window (90–110 ms), where significant differences in brain substrate activation were observed between all three types of stimuli. Such an early difference may reflect low-level encoding of tonal characteristics. Previously, it was found that both the brainstem, as indicated by frequency-following responses (FFR), and early ERP components of the cerebral cortex, such as N1, are sensitive to harmonic structure (Bidelman and Krishnan, 2009; Bidelman, 2013; Bianco et al., 2016; Kurt Steinmetzger, 2024). These early components occur automatically and indicate the processing of low-level acoustic features before conscious cognitive evaluation can take place. Our results partially support Bianco et al. (2016) finding, which demonstrated that consonant intervals evoke earlier and stronger responses in the auditory cortex compared to dissonant intervals, consistent with the perceptual superiority of consonance in early stages of processing.
In the 190–210 ms time window, dissonant chords evoked significantly broader activation than neutral stimuli, especially in bilateral posterior temporal and right frontocentral areas. This activation pattern corresponds to the P2 component (Schirmer and Kotz, 2006; Paulmann and Kotz, 2008; Schirmer, 2018). Dissonant stimuli typically evoke greater cognitive effort and affective engagement, and we suggest that the observed higher cortical activity may reflect this process. ERP studies have shown that significant differences are observed in this time window for pitch in musicians and the perception of roughness in non-musicians (Kung et al., 2014). That is, already at an early stage of processing, the acoustic properties of stimuli significantly influence their categorisation by a brain substrate.
In the 290–310 ms time window, dissonant stimuli evoked neural responses in the frontal, temporal, and occipital regions. This activity is associated with the P3 family component, which is typically linked to the automatic reorientation of attention to emotionally significant or novel stimuli. Previous studies of musical and emotional prosody have shown that incongruent or emotionally charged auditory stimuli reliably enhance the amplitude of P3 (Koelsch, 2010; Proverbio et al., 2020). The fact that presented musical stimuli evoked a pattern similar to affective vocal expressions supports the view that music and speech share common pathways for affective processing (Koelsch, 2014). These results suggest that dissonant chords, due to their affective significance, inherently attract attention and evoke an orienting response.
The ERP responses described above—characterised by well-defined N1, P2, and P3 components—closely resemble findings reported in the literature on emotional stimuli across multiple sensory modalities: emotional prosody and affective non-verbal sounds (e.g., screams, laughter, crying) elicit early modulation in components such as N1 and P2, underscoring the automatic and pre-attentive analysis of emotional value (Schirmer and Kotz, 2006; Paulmann and Kotz, 2008). Likewise, previous studies on emotional visual stimuli, such as negative, arousing pictures from standardised image sets such as the International Affective Picture System (IAPS) or EmoMadrid (Carretié et al., 2019), have reported an increase of early posterior negativity (EPN) at 200–300 ms and amplified late positive potentials (LPP), indicative of sustained attentional and evaluative processing (Olofsson et al., 2008; Hajcak et al., 2010; Carretié, 2014). This finding supports the hypothesis that partially shared neural mechanisms are involved in processing emotional salience in both music and language (Ramaswamy and Palaniswamy, 2024), in line with models of a shared prosodic-emotional network.
Neuroimaging studies reinforce the notion of overlapping emotional processing pathways for musical dissonance, emotional vocalisations, and affective images, with key limbic structures, such as the amygdala, consistently implicated across modalities (Koelsch, 2014; Frühholz et al., 2025). However, ERP topographies show differences specific to each modality, reflecting differential engagement of sensory-specific cortical regions (Olofsson et al., 2008; Filimonov et al., 2022; Proverbio et al., 2022). These findings align with Kölsch’s multilevel model of emotional processing of music, which posits that affective responses emerge through a hierarchy of neural processing phases, beginning with rapid subcortical evaluation and progressing gradually to more complex cortical evaluation (Koelsch, 2014). The initial ERP components observed in our study (e.g., N1 and P2) appear to represent automatic identification of affectively relevant features of the harmonic structure, consistent with subcortical processing. The later P3a, on the other hand, most likely represents higher-level cognitive appraisal and reorientation of attention to an affectively relevant stimulus. Thus, our ERP results support the hypothesis that emotional evaluation of musical consonances and dissonances is carried out through both hierarchically organised processes.
ERP components P2 and P3a, modulated by dissonant chords, resemble those observed in studies of emotional prosody and syntactic anomalies in speech (Friederici, 2011; Proverbio et al., 2020, 2022). This convergence lends support to the hypothesis of partially shared neural mechanisms for processing emotional salience in both music and language, a hypothesis that is consistent with models of a common prosodic-emotional network.
Our findings also agree with evidence of hemispheric asymmetries in music-induced emotion processing. Proverbio et al. (2020) demonstrated that consonant and dissonant chords activate different hemispheres—consonant sounds elicited stronger right temporal activation, whereas dissonant sounds preferentially activated left temporal regions. These asymmetries may reflect specialised processing of acoustic features (spectral vs. temporal resolution) and emotional valence in the auditory cortices.
We also found correlations between signal frequency characteristics and behavioural outcomes. The effects of the interaction between beta band amplitude (20–25 Hz) and dissonant stimuli indicate that high beta amplitude during stimulus presentation is associated with a significant reduction in the classification accuracy of dissonant chords. Specifically, a one-standard-deviation increase in beta amplitude corresponds to a 88% reduction in the odds of correctly identifying dissonant stimuli. This finding aligns with theories suggesting that beta oscillations are linked to the maintenance of current cognitive or perceptual sets and resistance to novel or deviant sensory input (Engel and Fries, 2010; Spitzer and Haegens, 2017). From this perspective, higher beta activity may indicate a perceptual condition in which the auditory system prioritises stability over flexibility, thereby reducing sensitivity to acoustically incongruent input, such as dissonance.
Two effects were observed for the gamma band amplitude (35–40 Hz). First, a main effect of stimulus type indicated that the probability of correctly identifying dissonant chords tended to be higher than that of consonant chords in the low gamma condition. However, this advantage appeared to be modulated by the amplitude of gamma: as gamma amplitude increased, the accuracy of detecting dissonant stimuli decreased markedly. In other words, dissonant chords were more discriminable at low gamma activity, but this advantage was weakened or reversed as gamma amplitude increased. Gamma oscillations have often been associated with attentional allocation, emotional salience, and perceptual binding (Başar-Eroglu et al., 1996; Mazaheri and Jensen, 2010). At the same time, excessive gamma activation may reflect heightened neural activity at the system level or network desynchronisation, particularly in response to emotional stimuli (Strube et al., 2021; Yu et al., 2022). Taken together, our results tentatively support the view that gamma activity contributes to stimulus evaluation but may reduce accuracy when over-amplified.
It should be noted that alpha band activity, often used in affective neuroscience as a metric for affective recovery and subjective preference, did not show a significant association with behavioural outcomes in our study. Although the alpha EEG is widely used in affective neuroscience due to its high reliability and reproducibility (Uusberg et al., 2013), its insensitivity in the present task might result from the sustained attention required during a relatively short (maximum running time of 16 min) randomised experimental session. Koelstra et al., for example, found a significant negative correlation between alpha EEG and valence, with higher arousal associated with lower alpha EEG amplitude when participants were exposed to music videos with varying emotional content. This interpretation is consistent with previous results indicating a decrease in alpha activity in task-focused conditions or during stimuli requiring evaluative processing (Koelstra et al., 2010; Paszkiel et al., 2020).
Several studies in the literature have shown a connection between neurodynamic EEG parameters and emotional responses (Choong et al., 2018). In our data, high-frequency LRTC (beta and gamma) showed associations primarily for trials categorised as “neutral.” However, the corresponding effect sizes were small in the beta range and unstable in the gamma range, suggesting that these findings should be interpreted with caution. One possible explanation is that neutral stimuli, unlike more emotionally charged ones, require more introspective processing, as reflected in both longer reaction times and lower accuracy. LRTC indices are believed to represent the balance between excitation and inhibition, as well as the temporal dynamics of information processing. These factors may be especially pertinent when participants try to classify a stimulus as neutral. Although preliminary, these findings suggest that LRTC may serve as a potential biomarker for distinguishing between neutral and affectively salient stimuli. However, this effect may also be related to the fact that the neutral type of stimulus was represented by the click of a metronome, the sound of which differed from the sound of the chords in timbre, as well as to the fact that the category “neutral” is less specific than “positive” or “negative.”
It is important to note several limitations of the study. First, the stimuli were synthesised using a single piano timbre. This ensured experimental control but limits generalisability, as timbre itself is a strong modulator of affective responses (Gabrielsson, 2001; Juslin and Västfjäll, 2008). Second, although the neutral condition served as a baseline, its acoustic structure (a single metronome click) was categorically different from consonant and dissonant chords. This choice ensured minimal affective salience, but it may also limit the direct comparability of the conditions and should be considered when interpreting the results. Third, the sample size and demographic homogeneity may restrict the generalisability of the findings. It should also be noted that all participants were non-musicians. Fourth, no preliminary registration was conducted to reduce the probability of Type I errors. Fifth, the odds ratio for some contrasts appeared numerically extreme with extensive CI. This is probably due to an imbalance in responses in some observations, rather than the actual size of the effect. Such inflated estimates are a known property of logistic models with rare events, and we therefore interpret them cautiously, focusing on the direction of the association rather than the absolute magnitude. Sixth, the use of categorical ratings (“pleasant,” “unpleasant,” “neutral”), rather than differentiated scales, limits the possibility of a more fine-grained interpretation of the effects. Whilst this approach ensured clear contrasts, it does not capture the full dimensional range of affective experience. Finally, the present findings were obtained in a Russian participant sample, and generalisation across cultures should be made with caution, given evidence that preferences for consonance and dissonance vary across populations (McDermott et al., 2016; Milne et al., 2023).
Importantly, although both consonant and dissonant chords were subjectively distinguished from neutral sounds, only dissonant chords were significantly distinguished at the level of EEG analysis. In the mid-latency (P2) and late (P3) intervals, dissonant chords showed widespread activation in frontal and temporoparietal regions, reflecting high affective salience associated with attention reorientation (Pallesen et al., 2005; Ellison et al., 2015). In contrast, consonant stimuli, although rated as pleasant at the behavioural level, failed to elicit comparably strong P2/P3 modulation under the same passive-listening conditions. This pattern is consistent with the negativity bias framework, according to which negatively valenced stimuli, across modalities, are prioritised in early perceptual and attentional processes because of their higher biological salience (Olofsson et al., 2008; Carretié, 2014), which partly explains the less striking neurophysiological response to positive valence in our data. At the same time, positive evaluations may involve later and more integrative processes, which are often modulated by cultural exposure and enculturation (McDermott et al., 2010; Argstatter, 2016). This temporal asymmetry may partly explain why consonant stimuli were behaviourally salient but less clearly differentiated in early ERP components. Whilst the unpleasant nature of dissonance is well-supported by multidisciplinary findings, the appealing characteristics of consonance apparently are the result of a complex interaction between cultural learning, memory, and specific cognitive and psychological mechanisms, in which culture plays a central role (Parncutt and Hair, 2018; Di Stefano and Spence, 2024). Thus, emotional valence in music is not merely a product of acoustic properties but deeply integrated with cognitive expectations and cultural backgrounds (Argstatter, 2016). For illustration, although infants universally demonstrate a preference for consonant intervals over dissonant ones (Trehub, 2003), certain isolated cultures exhibit no such preference for consonant chords (McDermott et al., 2010). This evidence suggests that culture has the capacity to modify innate preferences and elicit high-level emotional responses to harmonic structures. The “culture versus nature” debate has emphasised the interplay between psychoacoustic constraints (e.g., roughness and harmonicity) and enculturation (e.g., familiarity and context), as noted in complex music compositions, where dissonance can perform expressive or narrative functions without evoking unpleasant feelings (Juslin and Laukka, 2004). Key foundational studies advocate for composite models that integrate both sensory and cultural factors (Parncutt and Hair, 2018; see also Di Stefano and Spence, 2024, for recent re-examinations).
5 Conclusion
This experiment demonstrated how isolated piano timbre chords with varying harmonic structures elicit affective and neural responses. Behavioural outcomes demonstrated a reliable affective classification of consonant, dissonant, and neutral stimuli, confirming that musical consonance and dissonance correspond with subjective emotional valence.
Electrophysiological analysis revealed differential temporal properties and localisation of neural response sources between chord types. Dissonant and consonant chords elicited differential early activity in the temporal–parietal and auditory areas (90–110 ms), with the dissonant chords further displaying incised cortical activation at the middle latency (190–210 ms) and late (290–310 ms) time windows. Source localisation revealed the recruitment of parietal, frontal, and auditory areas, particularly in response to dissonant stimuli. Neural markers of perceptual sensitivity further complement these findings: elevated beta band amplitude (20–25 Hz) during the presentation of dissonant stimuli predicted significantly lower classification accuracy, suggesting reduced neural flexibility in processing incongruent input.
Additionally, increased gamma band amplitude (35–40 Hz) attenuated the behavioural advantage of dissonant chord recognition, highlighting a potential trade-off between salience-related gamma activation and perceptual precision. These findings suggest that musical harmony modulates both early perceptual and subsequent cognitive-affective stages in processing. Electrophysiological and behavioural responses to consonant and dissonant chords, as well as their comparison with neutral stimuli, support their potential usefulness as affective stimuli. However, further replication is required before chords can be considered established in this role. In addition to their potential utility for affective neuroscience, our findings underscore the promise of piano chords as stimuli for ERP research, particularly in cross-domain comparisons between music and language. Through their abstract form, chords can provide a controlled and versatile tool for investigating the general affective mechanisms underlying auditory communication.
Data availability statement
The datasets presented in this article are not readily available because the following investigations will be dependent on use of the data. Requests to access the datasets should be directed to Alexander Kirsanov, aXNicmFpbnR3aXN0ZXJAZ21haWwuY29t.
Ethics statement
The studies involving humans were approved by Ethics Committee of the Faculty of Psychology, Saint-Petersburg State University. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.
Author contributions
AK: Conceptualization, Formal analysis, Methodology, Project administration, Visualization, Writing – original draft, Writing – review & editing. MK: Data curation, Investigation, Writing – review & editing. DP: Data curation, Formal analysis, Visualization, Writing – review & editing. UN: Investigation, Software, Validation, Writing – review & editing. EB: Conceptualization, Methodology, Resources, Software, Supervision, Validation, Visualization, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. The research was supported by the grant of St. Petersburg State University 125022002751-7.
Acknowledgments
Authors acknowledge the Psychology Department of Saint-Petersburg State University. Authors are grateful to Elina Potsenkovskaia for the manuscript editing.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The authors declare that no Gen AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnhum.2025.1689067/full#supplementary-material
References
Allen, J. J. B., Coan, J. A., and Nazarian, M. (2004). Issues and assumptions on the road from raw signals to metrics of frontal EEG asymmetry in emotion. Biol. Psychol. 67, 183–218. doi: 10.1016/j.biopsycho.2004.03.007
Argstatter, H. (2016). Perception of basic emotions in music: culture-specific or multicultural? Psychol. Music 44, 674–690. doi: 10.1177/0305735615589214
Balconi, M., and Lucchiari, C. (2008). Consciousness and arousal effects on emotional face processing as revealed by brain oscillations. A gamma band analysis. Int. J. Psychophysiol. 67, 41–46. doi: 10.1016/j.ijpsycho.2007.10.002
Barr, D. J., Levy, R., Scheepers, C., and Tily, H. J. (2013). Random effects structure for confirmatory hypothesis testing: keep it maximal. J. Mem. Lang. 68, 255–278. doi: 10.1016/j.jml.2012.11.001
Başar-Eroglu, C., Strüber, D., Schürmann, M., Stadler, M., and Başar, E. (1996). Gamma-band responses in the brain: a short review of psychophysiological correlates and functional significance. Int. J. Psychophysiol. 24, 101–112. doi: 10.1016/S0167-8760(96)00051-7
Bianco, R., Novembre, G., Keller, P. E., Kim, S.-G., Scharf, F., Friederici, A. D., et al. (2016). Neural networks for harmonic structure in music perception and action. NeuroImage 142, 454–464. doi: 10.1016/j.neuroimage.2016.08.025
Bidelman, G. M. (2013). The role of the auditory brainstem in processing musically relevant pitch. Front. Psychol. 4:264. doi: 10.3389/fpsyg.2013.00264
Bidelman, G. M., and Grall, J. (2014). Functional organization for musical consonance and tonal pitch hierarchy in human auditory cortex. NeuroImage 101, 204–214. doi: 10.1016/j.neuroimage.2014.07.005
Bidelman, G. M., and Krishnan, A. (2009). Neural correlates of consonance, dissonance, and the hierarchy of musical pitch in the human brainstem. J. Neurosci. 29, 13165–13171. doi: 10.1523/JNEUROSCI.3900-09.2009
Boğa, M., Koyuncu, M., Kaça, G., and Bayazıt, T. O. (2023). Comparison of emotion elicitation methods: 3 methods, 3 emotions, 3 measures. Curr. Psychol. 42, 18670–18685. doi: 10.1007/s12144-022-02984-5
Bornas, X., Fiol-Veny, A., Balle, M., Morillas-Romero, A., and Tortella-Feliu, M. (2015). Long range temporal correlations in EEG oscillations of subclinically depressed individuals: their association with brooding and suppression. Cogn. Neurodyn. 9, 53–62. doi: 10.1007/s11571-014-9313-1
Bradley, M. M., and Lang, P. J. (2000). Affective reactions to acoustic stimuli. Psychophysiology 37, 204–215. doi: 10.1111/1469-8986.3720204
Brattico, E., Alluri, V., Bogert, B., Jacobsen, T., Vartiainen, N., Nieminen, S., et al. (2011). A functional MRI study of happy and sad emotions in music with and without lyrics. Front. Psychol. 2:308. doi: 10.3389/fpsyg.2011.00308
Brattico, E., Jacobsen, T., De Baene, W., Glerean, E., and Tervaniemi, M. (2010). Cognitive vs. affective listening modes and judgments of music – an ERP study. Biol. Psychol. 85, 393–409. doi: 10.1016/j.biopsycho.2010.08.014
Brown, S. B. R. E., van Steenbergen, H., Band, G. P. H., de Rover, M., and Nieuwenhuis, S. (2012). Functional significance of the emotion-related late positive potential. Front. Hum. Neurosci. 6:33. doi: 10.3389/fnhum.2012.00033
Carretié, L. (2014). Exogenous (automatic) attention to emotional stimuli: a review. Cogn. Affect. Behav. Neurosci. 14, 1228–1258. doi: 10.3758/s13415-014-0270-2
Carretié, L., Hinojosa, J. A., Martín-Loeches, M., Mercado, F., and Tapia, M. (2004). Automatic attention to emotional stimuli: neural correlates. Hum. Brain Mapp. 22, 290–299. doi: 10.1002/hbm.20037
Carretié, L., Tapia, M., López-Martín, S., and Albert, J. (2019). EmoMadrid: an emotional pictures database for affect research. Motiv. Emot. 43, 929–939. doi: 10.1007/s11031-019-09780-y
Choong, W. Y., Khairunizam, W., Omar, M. I., Murugappan, M., Abdullah, A. H., Ali, H., et al. (2018). EEG-based emotion assessment using Detrended Flunctuation analysis (DFA). J. Telecommun. Electron. Comput. Eng. JTEC 10, 105–109. Available online at: https://jtec.utem.edu.my/jtec/article/view/4132 (Accessed July 28, 2025).
Davidson, R. J. (1998). Anterior electrophysiological asymmetries, emotion, and depression: conceptual and methodological conundrums. Psychophysiology 35, 607–614. doi: 10.1017/S0048577298000134
Dawel, A., Miller, E. J., Horsburgh, A., and Ford, P. (2022). A systematic survey of face stimuli used in psychological research 2000–2020. Behav. Res. Methods 54, 1889–1901. doi: 10.3758/s13428-021-01705-3
Delplanque, S., Silvert, L., Hot, P., Rigoulot, S., and Sequeira, H. (2006). Arousal and valence effects on event-related P3a and P3b during emotional categorization. Int. J. Psychophysiol. 60, 315–322. doi: 10.1016/j.ijpsycho.2005.06.006
Di Stefano, N., Di Stefano, N., Vuust, P., Vuust, P., Brattico, E., and Brattico, E. (2022). Consonance and dissonance perception. A critical review of the historical sources, multidisciplinary findings, and main hypotheses. Phys Life Rev 43, 273–304. doi: 10.1016/j.plrev.2022.10.004
Di Stefano, N., and Spence, C. (2024). Should absolute pitch be considered as a unique kind of absolute sensory judgment in humans? A systematic and theoretical review of the literature. Cognition 249:105805. doi: 10.1016/j.cognition.2024.105805
Eerola, T., and Vuoskoski, J. K. (2011). A comparison of the discrete and dimensional models of emotion in music. Psychol. Music 39, 18–49. doi: 10.1177/0305735610362821
Ellison, D., Moisseinen, N., Fachner, J., and Brattico, E. (2015). Affective versus cognitive responses to musical chords: an ERP and behavioral study. Psychomusicol. Music Mind Brain 25, 423–434. doi: 10.1037/pmu0000127
Engel, A. K., and Fries, P. (2010). Beta-band oscillations — signalling the status quo? Curr. Opin. Neurobiol. 20, 156–165. doi: 10.1016/j.conb.2010.02.015
Ertl, M., Hildebrandt, M., Ourina, K., Leicht, G., and Mulert, C. (2013). Emotion regulation by cognitive reappraisal — the role of frontal theta oscillations. NeuroImage 81, 412–421. doi: 10.1016/j.neuroimage.2013.05.044
Filimonov, D., Railo, H., Revonsuo, A., and Koivisto, M. (2022). Modality-specific and modality-general electrophysiological correlates of visual and auditory awareness: evidence from a bimodal ERP experiment. Neuropsychologia 166:108154. doi: 10.1016/j.neuropsychologia.2022.108154
Fisher, R. A. (1992). “Statistical methods for research workers” in Breakthroughs in statistics. eds. S. Kotz and N. L. Johnson (New York, NY: Springer New York), 66–70.
Foo, F., King-Stephens, D., Weber, P., Laxer, K., Parvizi, J., and Knight, R. T. (2016). Differential processing of consonance and dissonance within the human superior temporal gyrus. Front. Hum. Neurosci. 10:154. doi: 10.3389/fnhum.2016.00154
Friederici, A. D. (2011). The brain basis of language processing: from structure to function. Physiol. Rev. 91, 1357–1392. doi: 10.1152/physrev.00006.2011
Friston, K. (2005). A theory of cortical responses. Philos. Trans. R. Soc. Lond. Ser. B Biol. Sci. 360, 815–836. doi: 10.1098/rstb.2005.1622
Fritz, T., Jentschke, S., Gosselin, N., Sammler, D., Peretz, I., Turner, R., et al. (2009). Universal recognition of three basic emotions in music. Curr. Biol. 19, 573–576. doi: 10.1016/j.cub.2009.02.058
Frühholz, S., Dietziker, J., Staib, M., Bobin, M., and Steiner, F. (2025). Evolutionary continuities and discontinuities in affective voice signaling. Emotion 25, 1079–1094. doi: 10.1037/emo0001484
Frühholz, S., Trost, W., and Grandjean, D. (2014). The role of the medial temporal limbic system in processing emotions in voice and music. Prog. Neurobiol. 123, 1–17. doi: 10.1016/j.pneurobio.2014.09.003
Gabrielsson, A. (2001). Emotion perceived and emotion felt: same or different? Musicae Sci. 5, 123–147. doi: 10.1177/10298649020050S105
Gerdes, A. B. M., Wieser, M. J., and Alpers, G. W. (2014). Emotional pictures and sounds: a review of multimodal interactions of emotion cues in multiple domains. Front. Psychol. 5:1351. doi: 10.3389/fpsyg.2014.01351
Güntekin, B., and Başar, E. (2014). A review of brain oscillations in perception of faces and emotional pictures. Neuropsychologia 58, 33–51. doi: 10.1016/j.neuropsychologia.2014.03.014
Hajcak, G., MacNamara, A., and Olvet, D. M. (2010). Event-related potentials, emotion, and emotion regulation: an integrative review. Dev. Neuropsychol. 35, 129–155. doi: 10.1080/87565640903526504
Harding, E., Kim, J. C., Demos, A., Roman, I., Tichko, P., Palmer, C., et al. (2025). Musical neurodynamics. Nat. Rev. Neurosci. 26, 293–307. doi: 10.1038/s41583-025-00915-4
Harrison, P. M. C., and Pearce, M. T. (2020). Simultaneous consonance in music perception and composition. Psychol. Rev. 27, 216–244. doi: 10.1037/rev0000169
Hensher, D. A., and Greene, W. H. (2003). The mixed logit model: the state of practice. Transportation 30, 133–176. doi: 10.1023/A:1022558715350
Horvat, M., Kukolja, D., and Ivanec, D. (2015). Comparing affective responses to standardized pictures and videos: A study report.
Hyvarinen, A. (1999). Fast ICA for noisy data using Gaussian moments, In: ISCAS’99. Proceedings of the 1999 IEEE International Symposium on Circuits and Systems VLSI (Cat. No.99CH36349), Orlando, FL: IEEE), 57–61.
Ishida, K., and Nittono, H. (2022). Relationship between early neural responses to syntactic and acoustic irregularities in music. Eur. J. Neurosci. 56, 6201–6214. doi: 10.1111/ejn.15856
Ito, T. A., Larsen, J. T., Smith, N. K., and Cacioppo, J. T. (1998). Negative information weighs more heavily on the brain: the negativity bias in evaluative categorizations. J. Pers. Soc. Psychol. 75, 887–900. doi: 10.1037/0022-3514.75.4.887
Juslin, P. N., and Laukka, P. (2004). Expression, perception, and induction of musical emotions: a review and a questionnaire study of everyday listening. J. New Music Res. 33, 217–238. doi: 10.1080/0929821042000317813
Juslin, P. N., and Västfjäll, D. (2008). Emotional responses to music: the need to consider underlying mechanisms. Behav. Brain Sci. 31, 559–575. doi: 10.1017/S0140525X08005293
Keil, A., Bradley, M. M., Hauk, O., Rockstroh, B., Elbert, T., and Lang, P. J. (2002). Large-scale neural correlates of affective picture processing. Psychophysiology 39, 641–649. doi: 10.1111/1469-8986.3950641
Knyazev, G. G. (2007). Motivation, emotion, and their inhibitory control mirrored in brain oscillations. Neurosci. Biobehav. Rev. 31, 377–395. doi: 10.1016/j.neubiorev.2006.10.004
Koelsch, S. (2010). Towards a neural basis of music-evoked emotions. Trends Cogn. Sci. 14, 131–137. doi: 10.1016/j.tics.2010.01.002
Koelsch, S. (2014). Brain correlates of music-evoked emotions. Nat. Rev. Neurosci. 15, 170–180. doi: 10.1038/nrn3666
Koelsch, S., Gunter, T., Friederici, A. D., and Schröger, E. (2000). Brain indices of music processing: “nonmusicians” are musical. J. Cogn. Neurosci. 12, 520–541. doi: 10.1162/089892900562183
Koelsch, S., Vuust, P., and Friston, K. (2019). Predictive processes and the peculiar case of music. Trends Cogn. Sci. 23, 63–77. doi: 10.1016/j.tics.2018.10.006
Koelstra, S., Yazdani, A., Soleymani, M., Mühl, C., Lee, J.-S., Nijholt, A., et al. (2010). “Single trial classification of EEG and peripheral physiological signals for recognition of emotions induced by music videos” in Brain Informatics. BI 2010. Lecture Notes in Computer Science. eds. Y. Yao, R. Sun, T. Poggio, J. Liu, N. Zhong, and J. Huang, vol. 6334 (Berlin, Heidelberg: Springer).
Krumhansl, C. L., and Jusczyk, P. W. (1990). Infants’ perception of phrase structure in music. Psychol. Sci. 1, 70–73. doi: 10.1111/j.1467-9280.1990.tb00070.x
Kruskal, W. H., and Wallis, W. A. (1952). Use of ranks in one-criterion variance analysis. J. Am. Stat. Assoc. 47, 583–621. doi: 10.1080/01621459.1952.10483441
Kumar, S., von Kriegstein, K., Friston, K., and Griffiths, T. D. (2012). Features versus feelings: dissociable representations of the acoustic features and valence of aversive sounds. J. Neurosci. 32, 14184–14192. doi: 10.1523/JNEUROSCI.1759-12.2012
Kung, C.-C., Hsieh, T.-H., Liou, J.-Y., Lin, K.-J., Shaw, F.-Z., and Liang, S.-F. (2014). Musicians and non-musicians’ different reliance of features in consonance perception: a behavioral and ERP study. Clin. Neurophysiol. 125, 971–978. doi: 10.1016/j.clinph.2013.10.016
Kurt Steinmetzger, A. R. (2024). The auditory P2 is influenced by pitch changes but not pitch strength and consists of two separate subcomponents. Imaging Neuroscience. MIT Press. Available online at: https://direct.mit.edu/imag/article/doi/10.1162/imag_a_00160/120751 (Accessed July 31, 2025).
Lang, P. J., and Bradley, M. M. (2010). Emotion and the motivational brain. Biol. Psychol. 84, 437–450. doi: 10.1016/j.biopsycho.2009.10.007
Li, M., and Lu, B.-L. (2009). Emotion classification based on gamma-band EEG. In: 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, IEEE, Minneapolis, MN, USA. 1223–1226.
Li, X., Song, D., Zhang, P., Zhang, Y., Hou, Y., and Hu, B. (2018). Exploring EEG features in cross-subject emotion recognition. Front. Neurosci. 12:162. doi: 10.3389/fnins.2018.00162
Lippman, E. A. (1994). A history of Western musical aesthetics. University of Nebraska Press, Lincoln, NE, USA.
Liu, X., Xu, Y., Alter, K., and Tuomainen, J. (2018). Emotional connotations of musical instrument timbre in comparison with emotional speech prosody: evidence from acoustics and event-related potentials. Front. Psychol. 9:737. doi: 10.3389/fpsyg.2018.00737
Maffei, A., and Angrilli, A. (2019). E-MOVIE - experimental MOVies for induction of emotions in neuroscience: an innovative film database with normative data and sex differences. PLoS One 14:e0223124. doi: 10.1371/journal.pone.0223124
Maris, E., and Oostenveld, R. (2007). Nonparametric statistical testing of EEG- and MEG-data. J. Neurosci. Methods 164, 177–190. doi: 10.1016/j.jneumeth.2007.03.024
Mazaheri, A., and Jensen, O. (2010). Rhythmic pulsing: linking ongoing brain activity with evoked responses. Front. Hum. Neurosci. 4:177. doi: 10.3389/fnhum.2010.00177
McDermott, J. H., Keebler, M. V., Micheyl, C., and Oxenham, A. J. (2010). Musical intervals and relative pitch: frequency resolution, not interval resolution, is special. J. Acoust. Soc. Am. 128, 1943–1951. doi: 10.1121/1.3478785
McDermott, J. H., Schultz, A. F., Undurraga, E. A., and Godoy, R. A. (2016). Indifference to dissonance in native Amazonians reveals cultural variation in music perception. Nature 535, 547–550. doi: 10.1038/nature18635
Milne, A. J., Smit, E. A., Sarvasy, H. S., and Dean, R. T. (2023). Evidence for a universal association of auditory roughness with musical stability. PLoS One 18:e0291642. doi: 10.1371/journal.pone.0291642
Miltner, W. H. R., Braun, C., Arnold, M., Witte, H., and Taub, E. (1999). Coherence of gamma-band EEG activity as a basis for associative learning. Nature 397, 434–436. doi: 10.1038/17126
Miskovic, V., and Schmidt, L. A. (2010). Cross-regional cortical synchronization during affective image viewing. Brain Res. 1362, 102–111. doi: 10.1016/j.brainres.2010.09.102
Nikulin, V. V., and Brismar, T. (2004). Long-range temporal correlations in alpha and beta oscillations: effect of arousal level and test–retest reliability. Clin. Neurophysiol. 115, 1896–1908. doi: 10.1016/j.clinph.2004.03.019
Nikulin, V. V., and Brismar, T. (2005). Long-range temporal correlations in electroencephalographic oscillations: relation to topography, frequency band, age and gender. Neuroscience 130, 549–558. doi: 10.1016/j.neuroscience.2004.10.007
Olofsson, J. K., Nordin, S., Sequeira, H., and Polich, J. (2008). Affective picture processing: an integrative review of ERP findings. Biol. Psychol. 77, 247–265. doi: 10.1016/j.biopsycho.2007.11.006
Pallesen, K. J., Brattico, E., Bailey, C., Korvenoja, A., Koivisto, J., Gjedde, A., et al. (2005). Emotion processing of major, minor, and dissonant chords: a functional magnetic resonance imaging study. Ann. N. Y. Acad. Sci. 1060, 450–453. doi: 10.1196/annals.1360.047
Parncutt, R., and Hair, G. (2018). A psychocultural theory of musical interval. Music. Percept. 35, 475–501. doi: 10.1525/mp.2018.35.4.475
Pascual-Marqui, R. D. (2002). Standardized low-resolution brain electromagnetic tomography (sLORETA): technical details. Methods Find Exp. Clin. Pharmacol. 24, 5–12.
Paszkiel, S., Dobrakowski, P., and Łysiak, A. (2020). The impact of different sounds on stress level in the context of EEG, cardiac measures and subjective stress level: a pilot study. Brain Sci. 10:728. doi: 10.3390/brainsci10100728
Patel, A. D. (2003). Language, music, syntax and the brain. Nat. Neurosci. 6, 674–681. doi: 10.1038/nn1082
Paulmann, S., and Kotz, S. A. (2008). Early emotional prosody perception based on different speaker voices. Neuroreport 19, 209–213. doi: 10.1097/WNR.0b013e3282f454db
Proverbio, A. M., De Benedetto, F., and Guazzone, M. (2020). Shared neural mechanisms for processing emotions in music and vocalizations. Eur. J. Neurosci. 51, 1987–2007. doi: 10.1111/ejn.14650
Proverbio, A. M., Tacchini, M., and Jiang, K. (2022). Event-related brain potential markers of visual and auditory perception: a useful tool for brain computer interface systems. Front. Behav. Neurosci. 16:1025870. doi: 10.3389/fnbeh.2022.1025870
Quiroga-Martinez, D. R., Basiński, K., Nasielski, J., Tillmann, B., Brattico, E., Cholvy, F., et al. (2022). Enhanced mismatch negativity in harmonic compared with inharmonic sounds. Eur. J. Neurosci. 56, 4583–4599. doi: 10.1111/ejn.15769
Ramaswamy, M., and Palaniswamy, S. (2024). Multimodal emotion recognition: a comprehensive review, trends, and challenges. Wiley Interdiscip. Rev Data Min. Knowl. Discov. 14:e1563. doi: 10.1002/widm.1563
Regnault, P., Bigand, E., and Besson, M. (2001). Different brain mechanisms mediate sensitivity to sensory consonance and harmonic context: evidence from auditory event-related brain potentials. J. Cogn. Neurosci. 13, 241–255. doi: 10.1162/089892901564298
Samson, A. C., Kreibig, S. D., Soderstrom, B., Wade, A. A., and Gross, J. J. (2016). Eliciting positive, negative and mixed emotional states: a film library for affective scientists. Cogn. Emot. 30, 827–856. doi: 10.1080/02699931.2015.1031089
Sangiuliano Intra, F., Avramiea, A.-E., Irrmischer, M., Poil, S.-S., Mansvelder, H. D., and Linkenkaer-Hansen, K. (2018). Long-range temporal correlations in alpha oscillations stabilize perception of ambiguous visual stimuli. Front. Hum. Neurosci. 12:159. doi: 10.3389/fnhum.2018.00159
Schad, D. J., Vasishth, S., Hohenstein, S., and Kliegl, R. (2020). How to capitalize on a priori contrasts in linear (mixed) models: a tutorial. J. Mem. Lang. 110:104038. doi: 10.1016/j.jml.2019.104038
Schirmer, A. (2018). Is the voice an auditory face? An ALE meta-analysis comparing vocal and facial emotion processing. Soc. Cogn. Affect. Neurosci. 13, 1–13. doi: 10.1093/scan/nsx142
Schirmer, A., and Kotz, S. A. (2006). Beyond the right hemisphere: brain mechanisms mediating vocal emotional processing. Trends Cogn. Sci. 10, 24–30. doi: 10.1016/j.tics.2005.11.009
Schubring, D., and Schupp, H. T. (2021). Emotion and brain oscillations: high arousal is associated with decreases in alpha- and lower Beta-Band power. Cereb. Cortex 31, 1597–1608. doi: 10.1093/cercor/bhaa312
Schweiger Gallo, I., Fernández-Dols, J.-M., Gollwitzer, P. M., and Keil, A. (2017). Grima: a distinct emotion concept? Front. Psychol. 8:131. doi: 10.3389/fpsyg.2017.00131
Song, C., Ikei, H., and Miyazaki, Y. (2021). Effects of forest-derived visual, auditory, and combined stimuli. Urban For. Urban Green. 64:127253. doi: 10.1016/j.ufug.2021.127253
Spearman, C. (1904). The proof and measurement of association between two things. Am. J. Psychol. 15:72. doi: 10.2307/1412159
Spitzer, B., and Haegens, S. (2017). Beyond the status quo: a role for Beta oscillations in endogenous content (re)activation. eNeuro 4, ENEURO.0170–ENEU17.2017. doi: 10.1523/ENEURO.0170-17.2017
Steinbeis, N., and Koelsch, S. (2011). Affective priming effects of musical sounds on the processing of word meaning. J. Cogn. Neurosci. 23, 604–621. doi: 10.1162/jocn.2009.21383
Strube, A., Rose, M., Fazeli, S., and Büchel, C. (2021). Alpha-to-beta- and gamma-band activity reflect predictive coding in affective visual processing. Sci. Rep. 11:23492. doi: 10.1038/s41598-021-02939-z
Sysoeva, O., Maximenko, V., Kuc, A., Voinova, V., Martynova, O., and Hramov, A. (2023). Abnormal spectral and scale-free properties of resting-state EEG in girls with Rett syndrome. Sci. Rep. 13:12932. doi: 10.1038/s41598-023-39398-7
Theódórsdóttir, D., and Höller, Y. (2023). EEG-correlates of emotional memory and seasonal symptoms. Appl. Sci. 13:9361. doi: 10.3390/app13169361
Trehub, S. E. (2003). The developmental origins of musicality. Nat. Neurosci. 6, 669–673. doi: 10.1038/nn1084
Uusberg, A., Uibo, H., Kreegipuu, K., and Allik, J. (2013). EEG alpha and cortical inhibition in affective attention. Int. J. Psychophysiol. 89, 26–36. doi: 10.1016/j.ijpsycho.2013.04.020
van der Vinne, N., Vollebregt, M. A., van Putten, M. J. A. M., and Arns, M. (2017). Frontal alpha asymmetry as a diagnostic marker in depression: fact or fiction? A meta-analysis. NeuroImage: Clinical 16, 79–87. doi: 10.1016/j.nicl.2017.07.006
Virtala, P., and Tervaniemi, M. (2017). Neurocognition of major-minor and consonance-dissonance. Music. Percept. 34, 387–404. doi: 10.1525/mp.2017.34.4.387
Vuust, P., Brattico, E., Seppänen, M., Näätänen, R., and Tervaniemi, M. (2012). The sound of music: differentiating musicians using a fast, musical multi-feature mismatch negativity paradigm. Neuropsychologia 50, 1432–1443. doi: 10.1016/j.neuropsychologia.2012.02.028
Wilcoxon, F. (1945). Individual comparisons by ranking methods. Biom. Bull. 1:80. doi: 10.2307/3001968
Yang, W., Makita, K., Nakao, T., Kanayama, N., Machizawa, M. G., Sasaoka, T., et al. (2018). Affective auditory stimulus database: an expanded version of the international affective digitized sounds (IADS-E). Behav. Res. Methods 50, 1415–1429. doi: 10.3758/s13428-018-1027-6
Yu, M., Shasha, X., Hua, M., Wang, H., Chen, X., Tian, F., et al. (2022). EEG-based emotion recognition in an immersive virtual reality environment: from local activity to brain network features. Biomed. Signal Process. Control 72:103349. doi: 10.1016/j.bspc.2021.103349
Keywords: chords, affective stimuli, ERP, source analysis, DFA, LRTC
Citation: Kirsanov A, Koriakina M, Panasenko D, Nikishkina U and Blagovechtchenski E (2025) Electrophysiological and behavioural responses to consonant and dissonant piano chords as standardised affective stimuli. Front. Hum. Neurosci. 19:1689067. doi: 10.3389/fnhum.2025.1689067
Edited by:
Susan Jerger, The University of Texas at Dallas, United StatesReviewed by:
Nicola Di Stefano, National Research Council (CNR), ItalyChristine Groß, Latvian Academy of Music, Latvia
Paola Crespo Bojorque, Universitat Pompeu Fabra, Spain
Copyright © 2025 Kirsanov, Koriakina, Panasenko, Nikishkina and Blagovechtchenski. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Alexander Kirsanov, aXNicmFpbnR3aXN0ZXJAZ21haWwuY29t
 Alexander Kirsanov1*
Alexander Kirsanov1*