ORIGINAL RESEARCH article

Front. Psychiatry

Sec. Computational Psychiatry

Volume 16 - 2025 | doi: 10.3389/fpsyt.2025.1633860

DGAT: a dynamic graph attention neural network framework for EEG emotion recognition

Provisionally accepted
  • 1Harbin Institute of Technology, Harbin, China
  • 2Shenzhen Academy of Aerospace Technology, Shenzhen, China

The final, formatted version of the article will be published soon.

Emotion recognition based on electroencephalogram (EEG) signals has shown increasing application potential in fields such as brain-computer interfaces and affective computing. However, current graph neural network models rely on predefined fixed adjacency matrices during training, which imposes certain limitations on the model's adaptability and feature expressiveness. In this study, we propose a novel EEG emotion recognition framework known as the Dynamic Graph Attention Network (DGAT). This framework dynamically learns the relationships between different channels by utilizing dynamic adjacency matrices and a multi-head attention mechanism, allowing for the parallel computation of multiple attention heads. This approach reduces reliance on specific adjacency structures while enabling the model to learn information in different subspaces, significantly improving the accuracy of emotion recognition from EEG signals. Experiments conducted on the EEG emotion datasets SEED and DEAP demonstrate that DGAT achieves higher emotion classification accuracy in both subjectdependent and subject-independent scenarios compared to other models. These results indicate that the proposed model effectively captures dynamic changes, thereby enhancing the accuracy and practicality of emotion recognition. DGAT holds significant academic and practical value in the analysis of emotional EEG signals and applications related to other physiological signal data.

Keywords: EEG, emotion recognition, Dynamic graph attention network, graph structure, Affective Computing

Received: 23 May 2025; Accepted: 23 Jun 2025.

Copyright: © 2025 Ding, Wang, Jiang, Xu, Bo, Ma and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Haifeng Li, Harbin Institute of Technology, Harbin, China

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.