- 1Shree L. R. Tiwari College of Engineering, Mumbai University, Mumbai, India
- 2Department of Information Technology, Mukesh Patel School of Technology Management and Engineering, SVKM’s Narsee Monjee Institute of Management Studies, Mumbai, India
- 3Terna Engineering College, Faculty of Science and Technology, University of Mumbai, Mumbai, India
- 4New Jersey Public Health Association, New Brunswick, NJ, United States
- 5NetScout Systems, Inc., Allen, TX, United States
Introduction: Identification and treatment of neurological disorders depend much on brain imaging and neurotherapeutic decision support. Although they are loud, do not remain in one spot, and are rather complex, electroencephalogram (EEG) signals are the principal tool used in research of brain function. This work employs an Adaptive Transformer-based technique with improved attention processes to extract temporal and spatial relationships in EEG data, effectively addressing these issues.
Methods: First processed to eliminate noise and split them into time-series chunks, EEG data are then included into the proposed approach. Channel-wise embeddings and temporal encoding help to depict the data. Then, a transformer design including spatial attention for inter-channel interactions, multi-head self-attention for temporal aspects, and an adaptive attention mask for domain-specific modifications is used. Other openly accessible EEG datasets as well as the TUH EEG Corpus and CHB-MIT were evaluated against the model. Its performance was scored using metrics like accuracy, precision, memory, and F1-score.
Results: The suggested method was more accurate than standard models like CNNs and LSTMs, with a score of 98.24%. The method was also shown to be able to find minor patterns in EEG data by improving precision and memory. Attention maps showed important areas of time and space, which made them easier to understand and useful in professional settings.
Discussion: The Adaptive Transformer turns out to be a useful tool for neurotherapeutic use of EEG data modeling. The approach provides greater medical assistance and knowledge on the functioning of the brain as well as answers significant issues. Future research might focus on subject-specific modifications and interaction with real-time systems.
Conclusion: This study demonstrates the potential of transformer-based models in revolutionizing EEG analysis for precision brain imaging and neurotherapeutic decision-making.
1 Introduction
Understanding brain function and detecting neurological diseases has long depend on electroencephalography (EEG), a fundamental instrument. EEG offers real-time understanding of brain dynamics by gathering electrical activity using non-invasive electrodes positioned on the head. Its uses are many and range from monitoring brain-computer interface systems to diagnosis of epilepsy and cognitive state evaluation. Because of its great temporal resolution, which makes it essential for research on the complex temporal and spatial patterns of the brain, EEG is crucial in precision brain imaging. Nevertheless, given the natural complexity of the signals, EEG analysis presents significant difficulties even with its promise (Ajali-Hernández et al., 2024; Dong et al., 2024).
Usually impacted by both internal and external events, EEG signals are famously non-stationary–that is, their statistical characteristics shift with time. This feature makes the identification of significant trends difficult as conventional approaches find difficulty to adjust to such dynamic changes. Another major obstacle is noise as ambient interference, muscle movements, or artifacts from eye blinks can contaminate EEG recordings. Furthermore very complicated temporal and spatial connections seen in EEG data need for sophisticated analysis methods to reveal underlying patterns. These difficulties have spurred a lot of study on more advanced computer approaches for EEG analysis (El Hadiri et al., 2024; Gunia et al., 2024).
Although conventional machine learning and deep learning models have been extensively used in EEG signal processing, they have natural restrictions. Support vector machines or random forests are two examples of classical machine learning methods that mostly depend on feature engineering, which calls for domain knowledge and usually misses the whole complexity of EEG signals. By automating feature extraction and developing hierarchical representations, deep learning models–CNN and LSTM among others–have shown a significant improvement (Ibitoye et al., 2021; Khayretdinova et al., 2024). These models, too, have limitations. CNNs are great in extracting spatial characteristics, but they often ignore the temporal dynamics that are very important in EEG analysis. On the other hand, LSTMs struggle with spatial connections between EEG channels while specializing in modeling sequential data. Usually requiring huge datasets to properly generalize, both methods are prone to overfitting when used on smaller, domain-specific datasets. Moreover, their interpretability is still restricted, which makes clinical adoption difficult when explainability is essential (Kim et al., 2024; Klooster et al., 2024).
With these constraints, the development of transformer-based models offers a transforming answer. Originally designed for “natural language processing,” (NLP), transformers with their attention techniques have transformed sequential data processing. Transformers are especially fit for complicated temporal and spatial modeling as they can record long-range relationships within data unlike conventional recurrent or convolutional architectures. Transformers’ fundamental invention is their self-attention mechanism, which gives input components dynamic weights according on their significance for the current job. This adaptability lets transformers concentrate on the most useful sections of the data, which fits very well with the difficulties of EEG signal analysis (Li et al., 2024). Transformers-based models have shown amazing success in different fields. Models like BERT and GPT have established new standards in NLP by deftly grasping semantics and context. Using both global and local information, transformers have outperformed conventional CNNs in computer vision applications like object identification and picture categorization. Time-series analysis, in which transformers have been used to anticipate financial patterns, track industrial processes, and project medical results, has likewise evolved from their usage. These developments show the adaptability and possibilities of transformers, therefore motivating their use in EEG analysis (Mutawa and Hassouneh, 2024).
Transformers’ capacity to solve important problems drives their acceptance for EEG study. Overcoming CNN and LSTM, their self-attention method can concurrently represent spatial linkages and temporal dynamics. Furthermore, transformers are naturally scalable and can handle big and complicated datasets free from sequential computation’s limitations (Dong et al., 2024; El Hadiri et al., 2024). For EEG research, where the data volume might be significant and real-time processing is usually needed, this scalability is very helpful. Moreover, transformers provide better interpretability than conventional deep learning models. Visualizing attention weights helps researchers and doctors understand which areas of the EEG data best support the predictions of the model, therefore strengthening confidence and enabling clinical decision-making.
This work aims to provide a unique adaptive transformer-based method for neurotherapeutic decision assistance and precision brain imaging. The following are our novel contributions:
• Using a domain-specific adaptive attention mask, the adaptive attention mechanism dynamically focuses on important temporal and spatial EEG characteristics for EEG variance. By stressing important areas and reducing noisy or pointless data, this approach guarantees consistency and resilience across subjects and situations. This approach targets significant patterns and reduces brainwave data variance, therefore enhancing the accuracy of EEG analysis.
• Integrating transformers for temporal-spatial modeling: Transformers endowed with “multi-head self-attention” record long-range temporal interdependence in EEG channels. Analyzing inter-channel interactions with spatial attention offers a whole temporal-spatial awareness of EEG data. This integration models concurrently time-based dynamics and spatial interdependencies, therefore enabling the whole interpretation and analysis of an EEG signal.
• Positional encoding provides an end-to-end framework to retain sequential structure of the EEG signal by means of feature embedding framework. Transformers convert raw EEG data into high-dimensional channel-wise embeddings, therefore effectively processing it. Through analysis of intricate temporal-spatial patterns in EEG data, this paradigm clarifies and guides application of brain activity dynamics.
This work offers many different kinds of contributions. It first presents a new framework using transformer architecture to replicate the complex spatial and temporal dynamics of EEG signals. Second, it visualizes attention weights to close the interpretability gap in EEG data and offers doctors practical insights. Third, it shows how scalable and flexible transformers are in managing various EEG datasets, hence opening the path for individualized neurotherapeutic uses. This work attempts to improve the accuracy, efficiency, and usefulness of EEG-based diagnoses and treatment by bridging the gap between modern computational approaches and clinical demands.
In EEG analysis for precision brain imaging and neurotherapeutic decision assistance, the suggested adaptive transformer-based method marks a major advance. This study introduces an adaptive transformer-based model that enhances EEG signal analysis by capturing complex spatial and temporal relationships. The proposed approach improves neurotherapeutic decision support and enables better precision in brain imaging. By integrating domain-specific modifications, it ensures robust and scalable performance across diverse EEG datasets. The combination of innovative models which include transformers with domain-specific knowledge offers excellent capability for commencing new horizons in neuroscience and clinical practice as the field develops.
2 Literature review
The study of brain dynamics through EEG analysis has become increasingly significant, offering a window into the functioning and disorders of the human brain. EEG provides a non-invasive, cost-effective approach to understanding brain activity, facilitating advancements in neurological diagnosis and therapeutic interventions. Despite its potential, the complexity of EEG signals, influenced by non-stationary characteristics and noise, presents substantial challenges for effective analysis. Advanced computational techniques, particularly those incorporating machine learning and deep learning, have been extensively employed to address these challenges, yet limitations remain.
Efforts to utilize self-supervised learning in EEG analysis have demonstrated promising outcomes, particularly in brain disease diagnosis, where adaptive node feature extraction has been shown to improve classification accuracy. This approach highlights the need for advanced modeling techniques capable of leveraging the hierarchical nature of brain networks (Zeng et al., 2024). Similarly, genetic variations influencing oscillatory alpha power in EEG data have been linked to specific neurobiological mechanisms, suggesting the potential for personalized models in precision brain imaging (Tichelman et al., 2023). The introduction of growth charts for brain function from infancy to adolescence provides a benchmark for understanding neural development, offering a reference point for the detection of anomalies (Iyer et al., 2024).
Deep learning-based frameworks have emerged as a dominant paradigm for EEG analysis, leveraging their ability to capture intricate temporal and spatial patterns. For example, research centered on classifying ADHD through deep learning methods have highlighted the role of particular mind regions inside the sickness, demonstrating the capability of neural networks in localizing pathophysiological markers (Sanchis et al., 2024). Also, reusable benchmarks for mind-age prediction have showcased the software of resting-state EEG signals in identifying age-associated neural changes, presenting a robust platform for exploring brain health throughout the lifespan (Engemann et al., 2022).
The hierarchical organization of complex correlation patterns in the brain has been effectively modeled to define a functional architecture, advancing the understanding of brain dynamics. Such models contribute to the identification of robust biomarkers for neurological conditions (Varga et al., 2024). Comparisons between EEG and MRI for early detection of “cortical dysmaturation” have highlighted the complementary nature of these modalities, emphasizing the need for integrative approaches in neuroimaging (White et al., 2024). The engineering of pluripotent stem cells for monitoring brain function and controlling neural activities further illustrates the interdisciplinary potential of EEG analysis (Cheng et al., 2024).
Sleep studies have also benefited from EEG biomarkers, particularly in understanding the activation of the brain’s lymphatic drainage system and its association with the blood-brain barrier. This opens avenues for exploring the therapeutic implications of sleep on neurological health (Semyachkina-Glushkovskaya et al., 2023). Meanwhile, transformer-based models have recently gained attention for their applicability in brain-computer interfaces, leveraging attention mechanisms to enhance the analysis of EEG signals (Pfeffer et al., 2024). Hybrid approaches combining EEG with other modalities like “functional near-infrared spectroscopy” (fNIRS) have further enriched the analytical capabilities, particularly in tasks requiring multidimensional feature extraction (Maher et al., 2023).
Scales Food and environmental elements have been shown to be sensitive by EEG studies on how outside events influence brain responses (Wu et al., 2022). Moreover, a novel approach linking visual stimuli to neural responses for cognitive analysis is stimulus-evoked EEG manifold learning for neural image classification (Falciglia et al., 2024). Finding indicators unique to the brain has demonstrated benefit from machine learning approaches. This makes more individualized therapies possible as well as more accurate diagnosis (Bomatter et al., 2024).
There are still issues even with these developments. Because EEG data is not always the same and has many dimensions, many times classic machine learning and deep learning models such CNNs and LSTM networks struggle with it. These techniques overfit, particularly in cases of noise presence, and need a lot of feature engineering to be done by hand. Furthermore, they are more difficult to grasp and use in different contexts as people are not always aware of the structural and geographical connections included within EEG patterns.
Originally designed for natural language processing applications, transformer-based models have developed a fresh approach of viewing EEG data. Long-range links and contextual linkages are quite well captured by these models. This qualifies them ideal for handling complicated in terms of time and spatial EEG data. Many fields have seen effective use of transformer designs as they can adapt to various data types and uncover interesting trends.
Building on these concepts, this work proposes to use a technique especially intended to operate with EEG data based on flexible transformers. This approach adds domain-specific modifications, such as a flexible attention mechanism, aiming to solve the issues with present models. This should make neurotherapeutic decision support and brain imaging more exact. This approach not only models complex data using the best characteristics of transformers but also provides us with a means to make EEG analysis more scalable and understandable. With this novel concept, the research aims to close the gaps in present approaches and therefore advance precision medicine and tailored brain care.
3 Methodology
As shown in the following Figure 1, The TUH EEG Corpus dataset was used in this study, providing a diverse collection of labeled EEG recordings that facilitate model evaluation across various neurological conditions, including seizure detection and cognitive state analysis. In continuation it included in this study are noise and non-stationarity’s related problems. The five basic steps that constitute the approach that has been described are preprocessing of the data, feature representation, transformer design, integration of adaptive attention masks, and the output layer customized for neurotherapeutic predictions. Every stage uses the use of mathematical ideas-based sophisticated methods in signal processing and machine learning.
3.1 Experimental setup
To assess the impact of hyperparameter variations, an ablation study was conducted by adjusting key parameters:
• Number of attention heads: Models were evaluated with 4, 8, and 12 heads, showing optimal performance at 8 heads.
• Model depth: 3-layer and 6-layer configurations were tested, with 6 layers achieving the best balance of accuracy and computational efficiency.
• Learning rate: A range from 0.0001 to 0.005 was explored, with 0.001 yielding optimal stability and convergence.
3.2 Dataset
The “TUH EEG Corpus” dataset was used in this study, providing a diverse collection of labeled EEG recordings that facilitate model evaluation across various neurological conditions, including seizure detection and cognitive state analysis. This dataset comprises labeled EEG signals spanning a broad spectrum of neurological illnesses, including cognitive assessments and seizure detection. Preprocessing the data helps to remove noise and artifacts, thereby ensuring the high quality of the input for the training and model testing. Their great number of temporal and spatial patterns makes them ideal for assessing novel transformer-based algorithms in neurotherapeutic uses.
3.3 Data preprocessing
Electroencephalogram signals are inherently noisy due to artifacts from muscle movement, eye blinks, and external interference. Preprocessing ensures that only relevant neural information is retained. The EEG data used in this study was recorded using the NeuroScan SynAmps2 EEG system with a 64-channel cap following the 10–20 electrode placement system, ensuring standardized spatial coverage. EEG signals were sampled at 256 Hz, and preprocessing included:
• Artifact Removal: techniques such as Independent Component Analysis (ICA) are employed to separate noise from neural signals as shown in Equation 1.
where X is the “observed EEG data”, A is the” mixing matrix”, and S is the “source signal matrix”. ICA estimates S by maximizing independence among components.
• Filtering: Band-pass filtering retains frequencies within a specific range (e.g., 0.5–50 Hz) used to remove low-frequency drifts and high-frequency noise as shown in Equation 2.
where H(f) is the “filter response” and X(f) is the “input signal” in the frequency domain.
• Segmentation: Signals are divided into “fixed-length time windows” (T) to capture temporal patterns, as shown in Equation 3, where t0 is the “start time of the window”, Signals were divided into 2-second epochs to capture relevant temporal patterns
3.4 Feature representation
To make EEG signals compatible with transformer models, they are converted into structured, high-dimensional embeddings as shown in Figure 2 and Table 1.
• Temporal Encoding: Position within the time-series sequence is encoded to retain temporal dependencies as shown in Equations 4, 5. A sinusoidal function is used
where pos is the “position”, i is the “dimension index” and d is the “embedding size”.
• Channel-wise embedding: Each EEG channel is treated as a separate sequence, and a linear embedding maps it to a higher-dimensional space, as shown in Equation 6.
3.5 Transformer architecture
The core of the proposed method is the transformer, which effectively models temporal-spatial relationships in EEG data. The proposed model is configured as follows:
• Layers: 6 Transformer encoder layers, balancing depth and computational efficiency.
• Attention heads: 8 multi-head attention units, ensuring effective feature extraction.
• Embedding size: 256 dimensions, capturing rich temporal-spatial representations.
• Feedforward network: Two fully connected layers with ReLU activation for non-linear transformations.
Ablation studies confirmed that increasing depth beyond 6 layers led to diminishing gains, while 8 attention heads provided the best trade-off between complexity and performance.
• Multi-head self-attention: Captures dependencies across time within each channel where Q,K,V are “query, key and value” matrices derived from input embeddings as shown in the following Equation 7
• Feedforward layers: Each attention output is processed through a position-wise feedforward network, as shown in Equation 8, where W1,W2 are weights and b1,b2 are the biases:
• Spatial attention: Captures inter-channel relationships as shown in Equation 9:
where As (i,j) represents the spatial attention weight between channel i and channel j, eij is the similarity score, and C is the total number of channels.
3.6 Adaptive attention mask
Electroencephalogram variability across subjects and conditions necessitates domain-specific adjustments. The adaptive attention mask emphasizes critical temporal-spatial regions dynamically.
• Dynamic masking: A learned mask M is applied to attention scores, as shown in Equation 10. Here A(i,j) is the original attention score, and M(i,j) is the adaptive mask weight.
• Optimization of mask: Masks are trained using a loss function that penalizes irrelevant regions as shown in Equation 11.
3.7 Output layer
The final layer converts the transformed embeddings into actionable predictions, either classification or regression outputs.
• Classification: A softmax function maps the output to probabilities for each class as shown in Equation 12,
where h is the hidden representation, Wk and bk are the weight and bias for class k, and K is the total number of classes.
The methodology concludes by integrating preprocessed EEG signals with feature extraction and adaptive transformer-based modeling, as depicted in Figure 3. This framework ensures precise temporal-spatial analysis and robust neurotherapeutic decision support.

Figure 3. Visualization of preprocessed EEG signal (2-second segment) and its power spectral density (PSD) distribution across frequencies.
The proposed Algorithm 1 for Adaptive Transformer-Based EEG Analysis leverages advanced transformer architectures to address the challenges of non-stationarity, noise, and temporal-spatial complexity inherent in EEG signals. By integrating domain-specific adaptive attention mechanisms, the algorithm dynamically prioritizes critical regions of the EEG data, enabling precise and interpretable neurotherapeutic predictions. This innovative approach ensures robust performance across diverse subjects and conditions.
3.8 Computation efficiency
While transformers are computationally intensive, optimizations ensure efficiency in real-time EEG analysis:
• Sparse attention mechanism: Reduces computation by focusing on relevant regions in the EEG sequence.
• Quantization & pruning: Lowers model size while maintaining accuracy, enabling deployment on edge devices.
• Sliding window processing: Instead of analyzing entire EEG sequences at once, real-time analysis is performed on overlapping time windows to minimize latency.
These techniques improve efficiency, enabling real-time inference without significantly sacrificing accuracy.
4 Results output and discussion
The adaptive transformer-based approach outperforms other EEG signal analysis methods for neurotherapeutic decision support as shown in Figure 4 and Table 2. The proposed method outperforms the decoding user’s movements method (97.33%), seizure prediction approach (94.6%), and interictal epileptiform discharge (IED) detection framework (95.2%) with 98.24% accuracy. This shows how transformers can capture complex temporal and spatial dependencies in EEG data.
At 97.8%, the proposed method has higher precision in identifying positive cases. Despite its 98.6% recall, the seizure prediction method’s 90.5% precision suggests a trade-off due to imbalanced datasets. The proposed method balances these metrics better, ensuring high sensitivity without sacrificing accuracy with 98.5% recall.
The proposed method’s 97.6% specificity in identifying negative cases shows its robustness. This is higher than all comparison methods, including seizure prediction (89.8%). The proposed method has the highest F1-score (98.15%), indicating balanced and reliable performance across diverse scenarios.
To evaluate robustness, EEG data with varying noise levels (10, 20, and 30% added Gaussian noise) was analyzed:
• With 10% noise, model accuracy remained above 97.5%, indicating strong resilience.
• At 20% noise, performance declined to 94.3%, reflecting moderate robustness.
• At 30% noise, accuracy dropped to 89.6%, suggesting sensitivity to extreme noise levels.
These findings indicate that effective artifact removal significantly enhances model reliability, emphasizing the need for robust preprocessing strategies.
The proposed approach has the highest area under the curve (AUC) value of 98.7%, supporting these findings. This metric highlights the method’s ability to distinguish classes across thresholds. Conclusions show that the proposed method can provide accurate, sensitive, and reliable neurotherapeutic decision support, overcoming limitations in existing methods.
5 Conclusion, limitation, future scope and implications
The suggested adjustable transformer-based technology is a useful method for evaluating EEG data for exact brain imaging and supporting medical professionals in making decisions about neurotherapeutic treatments. Temporal-spatial modeling, domain-specific flexible attention processes, and advanced feature representations help the method to be better than usual models in terms of accuracy, precision, and simplicity of understanding. This improvement has greatly helped to tackle some of the most important difficulties in EEG analysis–that of noise, non-stationarity, and complicated temporal-spatial correlations. In the framework of brain applications, it helps to uncover ideas with better accuracy and finally more usefulness.
Though the method has several good features, there are significant problems with it. Transformer models may be tricky to operate with in real time as they are difficult to compute, especially for big EEG datasets. Training depends on a lot of named data, hence it is likely that it will not function as expected with datasets lacking a lot of it. Though the domain-specific changes are useful, they must be optimized for every dataset. Sadly, this might cause their scalability to be less consistent under certain types of EEG settings.
Investigating lightweight transformer designs most effective for real-time applications and creating semi-supervised or unsupervised learning techniques that do not rely as much on labeled data can help one discover answers to these problems going forward. Investigating how this technique may be used with other techniques, including functional magnetic resonance imaging (MRI), would be interesting to help one to have a more complete knowledge of brain functioning.
The real world bears major consequences from this corpus of study. A strong basis is given which helps neurotherapeutic choices to be made with better accuracy, therefore enabling early evaluation and the formulation of a particular treatment plan. Apart from this, it might improve brain-computer interface technologies, which would lead to more use in cognitive neuroscience and therapy domains.
Data availability statement
The original contributions presented in this study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.
Author contributions
BN: Writing – original draft. VK: Writing – original draft. DS: Conceptualization, Writing – review & editing. KP: Writing – review & editing. SP: Writing – review & editing. UG: Investigation, Writing – original draft.
Funding
The author(s) declare that no financial support was received for the research and/or publication of this article.
Conflict of interest
SP was employed by NetScout Systems, Inc.
The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Ajali-Hernández, N., Travieso-González, C., Bermudo-Mora, N., Reino-Cacho, P., and Rodríguez-Saucedo, S. (2024). Study of an optimization tool avoided bias for brain-computer interfaces using a hybrid deep learning model. Irbm 45:100836. doi: 10.1016/j.irbm.2024.100836
Al-Quraishi, M., Elamvazuthi, I., Tang, T., Al-Qurishi, M., Adil, S., Ebrahim, M., et al. (2022). Decoding the user’s movements preparation from EEG signals using vision transformer architecture. IEEE Access. 10, 109446–109459. doi: 10.1109/ACCESS.2022.3213996
Bomatter, P., Paillard, J., Garces, P., Hipp, J., and Engemann, D. (2024). Machine learning of brain-specific biomarkers from EEG. eBioMedicine 106:105259. doi: 10.1016/j.ebiom.2024.105259
Cheng, J., Zhang, B., Cai, W., Zhao, S., Deng, X., Wang, B., et al. (2024). Engineering of rhesus monkey pluripotent stem cells for noninvasive survey and remote control after brain allotransplantation. Fundam Res. (in press). doi: 10.1016/j.fmre.2024.03.003
Dong, K., Zhang, L., Zhong, Y., Xu, T., Zhao, Y., Chen, S., et al. (2024). Meso-scale reorganization of local–global brain networks under mild sedation of propofol anesthesia. Neuroimage. 297:120744. doi: 10.1016/j.neuroimage.2024.120744
El Hadiri, A., Bahatti, L., El Magri, A., and Lajouad, R. (2024). Sleep stages detection based on analysis and optimisation of non-linear brain signal parameters. Results Eng. 23:102664. doi: 10.1016/j.rineng.2024.102664
Engemann, D., Mellot, A., Höchenberger, R., Banville, H., Sabbagh, D., Gemein, L., et al. (2022). A reusable benchmark of brain-age prediction from M/EEG resting-state signals. Neuroimage 262:119521. doi: 10.1016/j.neuroimage.2022.119521
Falciglia, S., Betello, F., Russo, S., and Napoli, C. (2024). Learning visual stimulus-evoked EEG manifold for neural image classification. Neurocomputing 588:127654. doi: 10.1016/j.neucom.2024.127654
Gunia, A., Moraresku, S., Janča, R., Ježdík, P., Kalina, A., Hammer, J., et al. (2024). The brain dynamics of visuospatial perspective-taking captured by intracranial EEG. Neuroimage 285:120487. doi: 10.1016/j.neuroimage.2023.120487
Ibitoye, R., Castro, P., Desowska, A., Cooke, J., Edwards, A., Guven, O., et al. (2021). Small vessel disease disrupts EEG postural brain networks in ‘unexplained dizziness in the elderly.’. Clin. Neurophysiol. 132, 2751–2762. doi: 10.1016/j.clinph.2021.07.027
Iyer, K. K., Roberts, J. A., Waak, M., Vogrin, S. J., Kevat, A., Chawla, J., et al. (2024). A growth chart of brain function from infancy to adolescence based on EEG. eBioMedicine 102:105061. doi: 10.1016/j.ebiom.2024.105061
Khayretdinova, M., Zakharov, I., Pshonkovskaya, P., Adamovich, T., Kiryasov, A., Zhdanov, A., et al. (2024). Prediction of brain sex from EEG: Using large-scale heterogeneous dataset for developing a highly accurate and interpretable ML model. Neuroimage 285:120495. doi: 10.1016/j.neuroimage.2023.120495
Kim, B., Ding, W., Yang, L., Chen, Q., Mao, J., Feng, G., et al. (2024). Simultaneous two-photon imaging and wireless EEG recording in mice. Heliyon 10:e25910. doi: 10.1016/j.heliyon.2024.e25910
Klooster, D., Voetterl, H., Baeken, C., and Arns, M. (2024). Evaluating robustness of brain stimulation biomarkers for depression: A systematic review of magnetic resonance imaging and electroencephalography studies. Biol. Psychiatry 95, 553–563. doi: 10.1016/j.biopsych.2023.09.009
Li, Z., Zhao, Y., Hu, Y., Li, Y., Zhang, K., and Gao, Z. (2024). Brain Stimulation Transcranial low-level laser stimulation in the near-infrared-II region (1064 nm) for brain safety in healthy humans. Brain Stimul. 17, 1307–1316. doi: 10.1016/j.brs.2024.11.010
Maher, A., Mian Qaisar, S., Salankar, N., Jiang, F., Tadeusiewicz, R., Pławiak, P., et al. (2023). Hybrid EEG-fNIRS brain-computer interface based on the non-linear features extraction and stacking ensemble learning. Biocybern. Biomed. Eng. 43, 463–475. doi: 10.1016/j.bbe.2023.05.001
Mutawa, A., and Hassouneh, A. (2024). Multimodal real-time patient emotion recognition system using facial expressions and brain EEG signals based on machine learning and log-sync methods. Biomed. Signal Process Control. 91:105942. doi: 10.1016/j.bspc.2023.105942
Pfeffer, M., Ling, S., and Wong, J. (2024). Exploring the frontier: Transformer-based models in EEG signal analysis for brain-computer interfaces. Comput. Biol. Med. 178:108705. doi: 10.1016/j.compbiomed.2024.108705
Rao, W., Zhang, L., Wang, X., Jiang, J., and Chen, D. (2024). in A Precise Interictal Epileptiform Discharge (IED) Detection Approach Based on Transformer BT - Applied Intelligence, eds D. Huang, P. Premaratne, and C. Yuan (Singapore: Springer Nature Singapore), 328–338.
Sanchis, J., García-Ponsoda, S., Teruel, M., Trujillo, J., and Song, I. Y. (2024). A novel approach to identify the brain regions that best classify ADHD by means of EEG and deep learning. Heliyon 10:e26028. doi: 10.1016/j.heliyon.2024.e26028
Semyachkina-Glushkovskaya, O. V., Karavaev, A., Prokhorov, M., Runnova, A., Borovkova, E., Ishbulatov, Y., et al. (2023). EEG biomarkers of activation of the lymphatic drainage system of the brain during sleep and opening of the blood-brain barrier. Comput. Struct. Biotechnol. J. 21, 758–768. doi: 10.1016/j.csbj.2022.12.019
Tichelman, N., Foerges, A., Elmenhorst, E., Lange, D., Hennecke, E., Baur, D., et al. (2023). A genetic variation in the adenosine A2A receptor gene contributes to variability in oscillatory alpha power in wake and sleep EEG and A1 adenosine receptor availability in the human brain. Neuroimage 280:120345. doi: 10.1016/j.neuroimage.2023.120345
Varga, L., Moca, V. V., Molnár, B., Perez-Cervera, L., Selim, M., Díaz-Parra, A., et al. (2024). Brain dynamics supported by a hierarchy of complex correlation patterns defining a robust functional architecture. Cell Syst. 15, 770–786.e5. doi: 10.1016/j.cels.2024.07.003
White, P., Ranasinghe, S., Chen, J., Van de Looij, Y., Sizonenko, S., Prasad, J., et al. (2024). Comparative utility of MRI and EEG for early detection of cortical dysmaturation after postnatal systemic inflammation in the neonatal rat. Brain Behav. Immun. 121, 104–118. doi: 10.1016/j.bbi.2024.07.028
Wu, B., Zhou, X., Blank, I., and Liu, Y. (2022). Investigating the influence of monosodium L-glutamate on brain responses via scalp-electroencephalogram (scalp-EEG). Food Sci. Hum. Wellness 11, 1233–1239. doi: 10.1016/j.fshw.2022.04.019
Zeng, Y., Lin, J., Li, Z., Xiao, Z., Wang, C., Ge, X., et al. (2024). Adaptive node feature extraction in graph-based neural networks for brain diseases diagnosis using self-supervised learning. Neuroimage 297:120750. doi: 10.1016/j.neuroimage.2024.120750
Keywords: EEG signal analysis, transformer architecture, neurotherapeutic decision support, temporal-spatial modeling, precision brain imaging, adaptive attention mechanism, machine learning in neurology
Citation: Nemade B, Kulkarni V, Shah D, Patel K, Patel S and Goradiya UB (2025) An adaptive transformer-based framework for advanced brain activity mapping and intelligent neurotherapeutic decision support. Front. Hum. Neurosci. 19:1551168. doi: 10.3389/fnhum.2025.1551168
Received: 24 December 2024; Accepted: 05 September 2025;
Published: 02 October 2025.
Edited by:
Mingzhou Ding, University of Florida, United StatesReviewed by:
Eddy Kwessi, Trinity University, United StatesMan Fai Leung, Anglia Ruskin University, United Kingdom
Copyright © 2025 Nemade, Kulkarni, Shah, Patel, Patel and Goradiya. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Vikram Kulkarni, dmlrcmFtLmt1bGthcm5pQG5taW1zLmVkdQ==