- Sichuan Provincial Maternity and Child Health Care Hospital, Chengdu, Sichuan, China
Introduction: The intricate interplay between endocrine systems and EEG signals is pivotal for understanding and managing physiological and neurological health. Traditional mathematical models often fail to capture the nonlinear dynamics, feedback mechanisms, and cross-system interactions inherent in these processes, limiting their applicability in clinical and research settings.
Methods: This study proposes a novel framework for modeling and analyzing the interaction between endocrine regulatory systems and EEG signals, leveraging advanced methodologies such as the Hormone Interaction Dynamics Network (HIDN) and the Adaptive Hormonal Regulation Strategy (AHRS). HIDN integrates graph-based neural architectures with recurrent dynamics to encapsulate the spatialtemporal interdependencies among endocrine glands, hormones, and EEG signal fluctuations. AHRS complements this by dynamically optimizing therapeutic interventions using real-time feedback and patient-specific parameters, ensuring adaptability to individual variability and external perturbations.
Results: The proposed model excels in scalability, precision, and robustness, addressing challenges like sparse clinical data, temporal resolution, and multi-hormonal regulation. Experimental validation demonstrates its efficacy in predicting hormone dynamics, EEG signal patterns, and therapeutic outcomes under varying conditions.
Discussion: This interdisciplinary approach bridges the gap between computational modeling and practical healthcare applications, advancing our understanding of endocrine-neurological interactions.
1 Introduction
The intricate interplay between the endocrine system and brain activity, as reflected in EEG signals Tao et al. (1), is a critical area of research for understanding physiological and pathological processes. This interaction not only provides insights into the neuroendocrine regulation of cognition, mood, and behavior but also enables the development of diagnostic and therapeutic tools for conditions such as hormonal imbalances, neurological disorders, and psychiatric diseases Cai et al. (2). Traditional approaches to analyzing these interactions have been limited in their ability to account for the nonlinear and dynamic nature of neuroendocrine mechanisms. Consequently, mathematical modeling has emerged as a promising framework to integrate diverse physiological processes, allowing researchers to decode complex feedback loops and predict system behaviors with greater precision. This study area, therefore, holds immense potential for advancing biomedical research and personalized medicine Li et al. (3).
A clearer understanding of the interaction between the endocrine system and brain activity is essential for advancing both theoretical knowledge and clinical practice. The endocrine system regulates a wide range of physiological functions—including metabolism, stress response, and circadian rhythms—through hormone release, which in turn influences brain function via modulation of neural excitability and signal transmission Kamble and Sengupta (4). Conversely, the brain exerts regulatory control over hormonal activity through neuroendocrine pathways, such as the hypothalamic-pituitary axes. Disruptions in this bidirectional communication are implicated in a variety of neurological and psychiatric disorders. Therefore, modeling these complex interdependencies holds the potential to improve early diagnosis, inform personalized treatment strategies, and enhance the efficacy of therapeutic interventions Pepino et al. (5). By bridging the gap between computational modeling and biological interpretation, our work aims to shed light on these dynamics and contribute to a more integrated understanding of human health.
Early approaches to mathematical modeling in this domain relied heavily on symbolic AI and knowledgebased systems Shen et al. (6). These models used differential equations and heuristic rules to represent the physiological processes governing endocrine systems and EEG signals. Techniques such as compartmental modeling were employed to capture hormone secretion, transport, and receptor interactions. Simultaneously, linear models were used to analyze EEG signal dynamics. Despite providing a structured framework and theoretical insights Song et al. (7), these methods suffered from oversimplification and limited accuracy in handling the inherent variability of biological system. Their inability to incorporate stochastic effects and adapt to real-time data further constrained their applicability, prompting the need for more robust methodologies Wang et al. (8).
To address the limitations of symbolic approaches, researchers turned to data-driven machine learning methods. These models utilized statistical techniques and algorithms Chudasama et al. (9), including regression analysis, support vector machines (SVM), and Bayesian networks, to identify correlations between endocrine parameters and EEG features Zhang et al. (10). Machine learning enabled the extraction of patterns from large datasets, improving prediction accuracy and adaptability to diverse conditions. However, these methods relied heavily on feature engineering and were limited by their dependence on labeled data Issa et al. (11). Additionally, machine learning models often failed to capture the temporal dependencies and feedback loops characteristic of neuroendocrine interactions, leaving room for improvement in representing the dynamic nature of these systems.
The advent of deep learning has significantly enhanced the modeling of endocrine-EEG interactions by leveraging advanced architectures such as recurrent neural networks (RNNs) Andayani et al. (12), long short-term memory networks (LSTMs), and transformers. These models excel in capturing temporal dependencies and nonlinear relationships within complex datasets, making them well-suited for analyzing time-series EEG signals and hormonal fluctuations Hu et al. (13). Moreover, the integration of pretrained models and attention mechanisms has improved their generalization and interpretability. However, the computational demands of deep learning, coupled with its black-box nature, pose challenges for scalability and clinical implementation. Additionally, the lack of physiological interpretability in purely data-driven models necessitates hybrid approaches that combine domain knowledge with advanced computation Dzedzickis et al. (14). The medical endocrine system is a complex network of glands and hormones responsible for regulating various physiological processes, including metabolism, growth, reproduction, and homeostasis. Understanding this system is essential for diagnosing and treating endocrine disorders, such as diabetes, thyroid dysfunction, and adrenal imbalances. The intricate interactions among glands, hormones, and target tissues present significant challenges for both clinical practice and computational modeling.
To validate the effectiveness of the proposed framework, we selected the emotion recognition task as the primary experimental setting. Emotion recognition from EEG signals offers a rich and well-established benchmark for evaluating models that aim to capture brain activity patternss Zhang et al. (15). More importantly, emotional states are known to be closely linked with hormonal fluctuations, making this task highly relevant for investigating endocrine-neurological interactions. By focusing on this domain, we are able to leverage existing annotated datasets and performance metrics while grounding our methodological contributions in a physiologically meaningful application Han et al. (16). This choice supports both the scientific relevance and practical utility of the proposed approach.
Given the aforementioned limitations, we propose a novel mathematical modeling framework that synergizes endocrine system dynamics and EEG signal analysis through hybrid methods. By integrating biologically informed differential equations with machine learning algorithms and leveraging deep learning for complex pattern recognition, our approach balances interpretability and accuracy. This model is designed to capture the bidirectional feedback between endocrine functions and brain activity, providing a comprehensive understanding of their interactions.
The proposed method has several key advantages:
● The proposed framework uniquely combines physiological modeling with deep learning to address the complexities of neuroendocrine interactions.
● It is versatile and efficient, enabling real-time analysis and application in diverse clinical and research scenarios.
● Empirical validation demonstrates superior accuracy in predicting hormonal influences on EEG patterns, enhancing diagnostic and therapeutic capabilities.
2 Related work
2.1 Endocrine system’s influence on neural activity
The endocrine system plays a crucial role in regulating brain function through the release of hormones that influence neural activity. Hormones such as cortisol, melatonin Shalbafan et al. (17), and thyroid hormones directly affect the amplitude and frequency of EEG signals, reflecting the state of the central nervous system Sarkar and Etemad (18). Mathematical modeling of these interactions has advanced significantly, employing differential equations to describe the dynamics of hormone release and their temporal effects on neural oscillations Kosti et al. (19). Recent research has focused on bidirectional feedback loops, where neural activity modulates hormonal levels and vice versa. These models often use nonlinear systems theory to account for the complexity of endocrine-neural interactions Li et al. (20). For instance, studies have developed models integrating circadian rhythm equations to predict EEG changes driven by hormonal fluctuations. Such approaches highlight the importance of capturing multi-scale dynamics, where endocrine processes operate on slower timescales compared to the rapid oscillations observed in EEG signals Shalbafan et al. (21). These models provide critical insights into disorders such as insomnia, depression, and hormonal imbalances, where disruptions in endocrine-neural coupling are prevalent Marini et al. (22).
2.2 EEG signal analysis in hormonal research
EEG signals provide a non-invasive method for assessing brain activity and are increasingly used to explore the effects of hormonal changes on neural function. Advanced signal processing techniques Liu et al. (23), such as time-frequency analysis, wavelet transforms, and independent component analysis Shalbafan et al. (24), have been applied to EEG data to identify biomarkers linked to endocrine activity. Mathematical models often use these biomarkers to establish causal relationships between hormone levels and EEG features. For example, spectral power in specific frequency bands, such as alpha and theta, has been correlated with hormonal states Lian et al. (25), enabling predictive models of endocrine influence. Furthermore, machine learning approaches, including support vector machines and neural networks Akhand et al. (26), have been employed to classify hormonal states based on EEG data. These methods often integrate mathematical frameworks, such as Bayesian inference or principal component analysis Pignatelli et al. (27), to enhance interpretability and accuracy. By combining EEG signal analysis with endocrine system modeling, researchers can construct comprehensive frameworks that elucidate the physiological underpinnings of neuroendocrine interactions Abbaschian et al. (28).
2.3 Hybrid models for neuroendocrine dynamics
Hybrid mathematical models combining endocrine system dynamics and EEG signal processing have emerged as powerful tools for investigating neuroendocrine interactions Heinonen et al. (29). These models typically integrate differential equations representing hormonal kinetics with computational methods for EEG signal interpretation Wani et al. (30). For example, coupled oscillatory models have been used to simulate the synchronization between hormonal cycles and neural rhythms observed in EEG data. Such models account for both intrinsic factors, like feedback regulation in hormone secretion Mehendale (31), and extrinsic factors, such as environmental stressors influencing neuroendocrine coupling. Moreover, hybrid frameworks often incorporate data-driven approaches, such as neural networks or optimization algorithms, to refine model parameters and improve predictive capabilities. These hybrid methods enable the exploration of complex phenomena, such as the interplay between chronic stress, cortisol levels, and EEG abnormalities Lv et al. (32). They also provide a foundation for developing personalized interventions by simulating individual-specific neuroendocrine responses Mo et al. (33). By bridging mechanistic and statistical approaches, hybrid models represent a comprehensive strategy for advancing the understanding of endocrine and EEG signal interactions Islam et al. (34).
3 Method
3.1 Overview
This paper proposes a novel framework to model and forecast endocrine system dynamics, leveraging advanced computational approaches to address the limitations of existing methodologies. This section provides an overview of our contributions and outlines the focus of the subsequent sections.3.2 introduces the Preliminaries, offering a mathematical formalization of the endocrine system’s regulatory mechanisms. This section explores the system’s feedback loops, hormone secretion patterns, and interdependencies, forming the basis for the proposed modeling framework.3.3 presents our Innovative Endocrine Modeling Framework, designed to capture the system’s dynamic behaviors with high fidelity. This model integrates physiological insights with data-driven methodologies, enabling accurate simulation and prediction of hormone fluctuations under various conditions.3.4 details the Adaptive Intervention Strategy, highlighting how domain-specific adaptations are incorporated into the framework to optimize therapeutic interventions. This strategy addresses the variability in individual responses and the challenges posed by incomplete or noisy clinical data.Together, these components establish a comprehensive approach to studying the endocrine system, bridging the gap between theoretical understanding and practical applications. The subsequent sections provide an in-depth exploration of the theoretical constructs, algorithmic innovations, and empirical validations underlying our framework.
While our framework builds upon foundational elements in endocrine modeling and deep learning, several components represent distinct methodological innovations. The use of graph-based neural architectures to represent inter-glandular hormone dynamics draws on established graph neural network principles; however, the specific design of the Hormone Interaction Dynamics Network (HIDN)—which combines attentionmodulated graph updates with LSTM-based temporal modeling and external stimulus encoding—is novel and tailored for endocrine-EEG integration. The Adaptive Hormonal Regulation Strategy (AHRS) introduces a real-time feedback mechanism with dynamic optimization and risk-aware personalization, which, to our knowledge, has not been previously applied in this context. These innovations allow our model to more effectively simulate nonlinear, multi-hormonal interactions and adapt interventions to individual physiological states, setting it apart from existing approaches.
3.2 Preliminaries
The endocrine system is a network of glands that release hormones to regulate physiological processes. This section formalizes the system’s dynamics and interactions through mathematical constructs to facilitate predictive modeling and analysis. Hormonal regulation involves complex feedback mechanisms, crossgland interactions, and temporal dependencies, all of which must be represented within a cohesive framework.
Let denote the set of endocrine glands, where is the total number of glands. Each gland secretes a hormone , which influences a target organ or another gland.The concentration of hormone at time is denoted by , forming the state vector . The rate of change of is governed by Equation 1:
where represents the regulatory dynamics, encapsulates external stimuli or interventions, and is a vector of physiological parameters.
Hormonal regulation is often mediated by negative feedback loops, ensuring homeostasis. For example, consider the hypothalamic-pituitary-thyroid (HPT) axis. The hypothalamus secretes thyrotropin-releasing hormone (TRH), which stimulates the pituitary to produce thyroid-stimulating hormone (TSH). TSH, in turn, prompts the thyroid to release thyroxine () and triiodothyronine (). The feedback relationship is modeled as Equation 2:
where is the feedback sensitivity coefficient, and represents the concentration of . Similar constructs apply to other axes such as the hypothalamic-pituitary-adrenal (HPA) and hypothalamic-pituitary-gonadal (HPG) systems.
Interactions between glands result in coupled differential equations. For example, the influence of gland on gland can be represented by Equation 3:
where is the coupling strength, (·) captures the secretion response, and is the natural degradation rate of .
External interventions, such as drug administration or environmental changes, are modeled as inputs . For instance, the administration of insulin to regulate glucose levels in diabetes can be expressed as Equation 4:
where is the administered dose and is the administration interval.
The complete system dynamics are represented as Equation 5:
where F is the vector-valued function describing the combined effects of feedback, interactions, and external stimuli. Solving these equations requires numerical methods due to their nonlinearity and interdependencies.
The physiological parameters p vary across individuals and are often difficult to measure directly. Hormonal dynamics operate on multiple timescales, necessitating adaptive methods for capturing both short-term fluctuations and long-term trends.Clinical measurements of hormone levels are infrequent, complicating model calibration.
3.3 Hormone interaction dynamics network
In this section, we introduce the Hormone Interaction Dynamics Network (HIDN), a cutting-edge model designed to capture intricate, multi-layered interactions within the endocrine system (Algorithm 1). By integrating domain-specific physiological knowledge with advanced neural architectures, HIDN models the dynamics of hormone secretion, regulation, and inter-gland interactions with precision (As shown in Figure 1).

Figure 1. Hormone Interaction Dynamics Network (HIDN) is a comprehensive framework, combining Graph Neural Networks (GNNs), Long Short-Term Memory (LSTM) networks, and external stimuli encoding to model hormone secretion, regulation, and inter-gland interactions. The model leverages graph-based representations to capture spatial dependencies among glands and employs recurrent neural structures to model temporal hormone dynamics. Integration of external factors, such as drug administration or environmental changes, is achieved through encoding mechanisms. Missing-signal prompts and generative prompts further enhance the model’s robustness in handling incomplete data. The overall system architecture ensures precise prediction of hormone concentration trends and adaptive response modeling in endocrine systems.
Algorithm 1. Hormone Interaction Dynamics Network (HIDN).
3.3.1 Graph-Based Representation for Inter-Gland Dynamics
HIDN models the endocrine system as a directed graph , where nodes represent glands and edges capture regulatory influences between glands. The state of each node is defined by its hormone concentration at time , while edge weights quantify the influence of gland on gland . This graph structure enables a detailed representation of the system’s spatial dependencies. The dynamics of node states are updated iteratively using a Graph Neural Network (GNN), formulated as Equation 6:
where denotes the neighborhood of node , are attention weights learned to prioritize influential connections, are layer-specific trainable weight matrices, and is a non-linear activation function such as ReLU. The attention weights are computed via Equation 7:
where is a learnable parameter vector and denotes concatenation. This mechanism adaptively assigns importance to each connection based on the dynamic state of the graph.
The update rule for the entire graph can be compactly expressed as Equation 8:
where is the matrix of node states at layer , is the attention-modulated adjacency matrix, and are trainable weights. To capture the feedback mechanisms inherent in endocrine systems, HIDN incorporates self-loops in , represented by Equation 9:
where is the identity matrix and is the original attention-modulated adjacency matrix.
To ensure stability in hormone concentration predictions, a normalization step is applied to node states Equation 10:
where represents the L2 norm. This normalization prevents amplification of hormone concentrations and ensures consistent updates across iterations.
Finally, the overall node update process iterates across layers, generating hierarchical representations that capture increasingly complex inter-gland dependencies Equation 11:
This comprehensive graph-based formulation allows HIDN to accurately model the regulatory mechanisms and feedback loops critical to endocrine system dynamics.
3.3.2 Temporal Dynamics with Recurrent Neural Structures
To accurately model the temporal evolution of hormone concentrations, HIDN employs Long Short-Term Memory (LSTM) networks, which are designed to capture both short-term and long-term dependencies in sequential data (As shown in Figure 2). Each gland maintains a hidden state and a cell state , which are updated at each time step based on the current hormone concentration and the previous states Equations 12–15:

Figure 2. The diagram illustrates Temporal Dynamics with Recurrent Neural Structures in HIDN, leveraging Long Short-Term Memory (LSTM) networks to accurately model both short-term and long-term temporal evolution and dependencies of hormone concentrations. The architecture integrates task and output tokens through cross-attention mechanisms, followed by multilayer perceptron (MLP) layers and linear transformations, effectively encoding complex temporal interactions. Image embeddings and mask adapters refine the model’s predictions, enabling HIDN to capture intricate feedback loops, delayed responses, and adaptive temporal trends inherent within the endocrine system dynamics. This comprehensive recurrent framework significantly enhances the accuracy and robustness of hormone concentration predictions over time.
Here, , , and are the forget, input, and output gates, respectively, which control the flow of information within the LSTM. The parameters , , and are learnable weights and biases.
The cell state is updated as Equation 16:
where denotes element-wise multiplication. The hidden state is then computed as Equation 17:
The combined hidden states of all glands form the latent representation Equation 18:
where is the number of glands. This representation captures the temporal dependencies and interactions across the endocrine system.
To incorporate external inputs , such as interventions or environmental changes, the LSTM update is modified as Equation 19:
where is an input encoding function.
The predicted hormone concentrations for the next time step are obtained as Equation 20:
where ω maps the hidden states to the predicted concentrations. This recurrent framework allows HIDN to model feedback loops, delayed responses, and adaptive temporal trends in hormone dynamics with high accuracy.
3.3.3 Integration of External Stimuli through Encoding
To capture the impact of external factors such as drug administration, environmental changes, or physiological interventions, HIDN incorporates an encoding mechanism that transforms these stimuli into representations compatible with the network’s internaldynamics. Let denote the external inputs at time . These inputs are processed through a multi-layer perceptron (MLP) (·), parameterized by Equation 21:
where (·) consists of multiple linear transformations interleaved with activation functions such as ReLU Equation 22:
with , , , and as learnable parameters, and denoting the activation function.
The encoded external stimuli are integrated into the LSTM-based temporal dynamics. For each gland , the hidden state evolves based on prior states , current hormone concentrations , and the encoded input Equation 23:
where is the LSTM update function parameterized by . The temporal updates incorporate both intrinsic gland dynamics and external influences, enabling HIDN to model adaptive responses.
The combined hidden states for all glands at time form the latent representation Equation 24:
where is the number of glands.
The predicted hormone concentrations for the next time step are computed as Equation 25:
where (·) is a mapping function implemented as another MLP, parameterized by Equation 26:
with and as trainable parameters.
The training objective minimizes the mean squared error (MSE) between predicted and observed hormone concentrations Equation 27:
where Θ includes all learnable parameters in the model.
3.4 Adaptive hormonal regulation strategy
The Adaptive Hormonal Regulation Strategy (AHRS) leverages the predictive capabilities of the Hormone Interaction Dynamics Network (HIDN) to design personalized interventions for managing endocrine disorders (Algorithm 2). AHRS dynamically adjusts therapeutic strategies in response to real-time physiological changes, optimizing treatment outcomes while minimizing adverse effects (As shown in Figure 3). The core innovations of AHRS are outlined below:

Figure 3. Overview of the Adaptive Hormonal Regulation Strategy (AHRS), illustrating the integration of real-time hormonal monitoring and predictive feedback mechanisms, personalized risk-aware adaptation using bidirectional Mamba modules for pixel-level and feature-map analysis, and dynamic multi-objective optimization for intervention design to balance therapeutic efficacy and safety. The strategy continuously updates interventions based on hormone dynamics predictions and observed physiological responses, ensuring optimized, individualized endocrine management. The schematic highlights interactions among system components, emphasizing AHRS’s adaptability, predictive accuracy, and risk mitigation capabilities.
Algorithm 2. Adaptive Hormonal Regulation Strategy (AHRS).
3.4.1 Real-Time Hormonal Monitoring and Predictive Feedback
The Adaptive Hormonal Regulation System (AHRS) integrates continuous real-time monitoring of hormone levels to enable dynamic and personalized feedback mechanisms. Let represent the observed hormonal concentrations at time t. These real-time observations are fed into HIDN, which predicts the future hormonal dynamics based on current states, external interventions, and patient-specific parameters Equation 28:
where u(t) denotes external interventions applied at time t, such as medication or environmental changes, and p encapsulates physiological parameters specific to the individual.
To assess the system’s accuracy, AHRS calculates the prediction error, or deviation, as Equation 29:
This deviation reflects discrepancies between predicted and observed hormone levels, which may result from unmodeled external factors or intrinsic variability in the endocrine system.
A feedback loss function is defined to quantify and minimize this deviation Equation 30:
where denotes the squared Euclidean norm. The objective of the feedback mechanism is to iteratively reduce by adjusting therapeutic strategies.
The dynamic adjustment of interventions is achieved through an optimization step Equation 31:
where is the learning rate for intervention adjustments. This gradient-based update ensures that future interventions align closely with the predicted physiological trajectory, minimizing deviations.
To account for real-time changes in patient states, AHRS integrates the updated interventions back into the predictive model Equation 32:
allowing iterative refinement of predictions and treatments.
The system also evaluates the stability of predictions and feedback by monitoring the convergence of deviations over time Equation 33:
A stability index near 1 indicates consistent performance, while significant fluctuations trigger recalibration of HIDN or external interventions.
3.4.2 Dynamic Multi-Objective Optimization for Intervention Design
The Adaptive Hormonal Regulation System (AHRS) utilizes a dynamic multi-objective optimization framework to design therapeutic interventions that regulate interacting hormone levels while minimizing associated risks (As shown in Figure 4). For a system of N hormones, the optimization is guided by the following composite loss function Equation 34:

Figure 4. Overview of Dynamic Multi-Objective Optimization for Intervention Design in AHRS. This framework leverages advanced neural network architectures to optimize therapeutic interventions that regulate hormone levels while minimizing associated risks. The system architecture combines upsampling, patch embedding, transformer blocks, and LoRA layers for enhanced feature extraction and representation learning. An attention module integrates CBAM and ECA mechanisms to capture complex interactions, while the decoder reconstructs signals for analysis. The optimization process utilizes a composite loss function balancing target hormone regulation with safety considerations, incorporating gradient-based adjustments to minimize deviations and adapt to dynamic hormonal changes. This robust approach ensures precise and adaptive therapeutic interventions for hormonal regulation.
where: - represents the observed level of hormone at time , - is the desired target level for hormone , - is a weight reflecting the clinical importance of maintaining hormone at its target level, - is a risk function quantifying potential adverse effects of the intervention , - is a regularization parameter balancing therapeutic objectives and safety.
The risk function is modeled to penalize excessive or conflicting interventions Equation 35:
where penalizes the overall intensity of interventions, and reflects the risk of interaction between interventions targeting hormones and .
Interventions are iteratively adjusted to minimize . At each time step , the update rule is defined as Equation 36:
where η is the learning rate controlling the step size for updates, and is the gradient of the loss with respect to the intervention u(t).
The gradient is computed as Equation 37:
where captures the sensitivity of hormone i to the intervention u, and is the gradient of the risk function.
To ensure convergence and stability, AHRS monitors the total deviation from target levels Equation 38:
and adjusts the learning rate dynamically based on Equation 39:
where is a decay factor. By iteratively refining interventions, AHRS ensures precise regulation of hormone levels while minimizing risks, providing a robust and safe framework for dynamic therapeutic optimization.
3.4.3 Personalization and Risk-Aware Adaptation
To accommodate individual variability in endocrine dynamics, AHRS refines patient-specific parameters using Bayesian inference. This probabilistic framework integrates observed patient data to update the posterior distribution of Equation 40:
where is the likelihood of the data given the parameters, and is the prior distribution reflecting prior knowledge about the patient’s physiological state. This process enables AHRS to personalize interventions by iteratively refining as more data becomes available.
To ensure that therapeutic interventions remain safe, AHRS integrates a risk-aware constraint mechanism. Let represent a set of clinical safety constraints applied to the intervention u(t). AHRS optimizes the intervention as Equation 41:
where represents the allowable risk threshold. The safety constraints may include limits on dosage intensity, interaction risks, or patient-specific contraindications Equation 42:
where penalizes conflicting interventions targeting hormones and .
The optimization process employs a Lagrangian formulation to incorporate these constraints Equation 43:
where is a penalty parameter that enforces adherence to safety thresholds.
Real-time feedback further enhances personalization by dynamically adapting interventions based on observed deviations. The patient-specific update rule is given by Equation 44:
where is the learning rate for parameter adaptation. This ensures that AHRS continually aligns its model to the patient’s evolving physiological state.
The adjusted interventions are integrated into the predictive feedback loop Equation 45:
allowing for precise and personalized predictions of hormone dynamics. To address the difficulty of directly measuring individual-specific physiological parameters p, AHRS incorporates a Bayesian inference mechanism that dynamically estimates these latent variables based on observed hormone trajectories and external interventions. Rather than relying on fixed parameter values, the model initializes a prior distribution informed by physiological norms, which is subsequently updated using observed hormonal responses through the posterior , where D denotes the set of hormone measurements and intervention histories. This adaptive estimation allows AHRS to personalize regulation strategies for each individual by inferring feedback sensitivities, degradation rates, and regulatory couplings without requiring direct clinical measurement. The inferred parameters are integrated into the predictive loop of HIDN, enabling precise hormone trajectory forecasting and real-time intervention adjustments tailored to the individual’s physiological profile.
4 Experimental setup
4.1 Dataset
The PhyAAt Dataset Ahuja and Setia (35) is a comprehensive resource designed for studying physiological responses and affective states. It contains multimodal data, including heart rate, galvanic skin response, and electroencephalography (EEG) signals, collected from participants under various controlled emotional stimuli. This dataset is widely used for emotion recognition, stress analysis, and human-computer interaction studies due to its rich annotations and diverse range of emotional scenarios, enabling robust evaluation of affective computing models. The Physionet MI Dataset Hammad et al. (36) is a benchmark dataset for motor imagery (MI) tasks, comprising EEG recordings collected from subjects performing imaginary hand and foot movements. It includes well-structured signals with detailed metadata such as trial annotations and channel configurations. This dataset is highly valuable for developing and benchmarking brain-computer interface (BCI) systems, facilitating advancements in motor rehabilitation and neurofeedback applications through its standardized and reproducible experimental setup. The BIDS Siena Scalp EEG Dataset Dan et al. (37) adheres to the Brain Imaging Data Structure (BIDS) standard, featuring high-resolution scalp EEG recordings from multiple participants. This dataset provides a well-organized framework for studying brain dynamics and neurological disorders such as epilepsy and Alzheimer’s disease. Its uniform data structure, combined with metadata annotations, makes it suitable for machine learning and deep learning models aimed at brain signal analysis and clinical applications. The STEW Dataset Siddhad et al. (38) is a spatio-temporal EEG dataset designed for analyzing stress and workload in human participants. It contains multi-channel EEG recordings collected during task performance under varying levels of cognitive load. This dataset is instrumental in understanding brain dynamics related to stress and workload, with applications in ergonomics, workplace efficiency, and mental health monitoring. Its extensive labeling and high temporal resolution make it an essential benchmark for exploring cognitive and emotional states through EEG signal analysis.
4.2 Experimental details
The proposed model was evaluated on four datasets: PhyAAt, Physionet MI, BIDS Siena Scalp EEG, and STEW. The implementation utilized PyTorch 2.0, with experiments conducted on NVIDIA A100 GPUs featuring 40 GB of VRAM. Model parameters were initialized using the Xavier method, and the Adam optimizer was used with an initial learning rate of 0.0005, reduced by a factor of 0.85 every 10 epochs. A total of 150 epochs were used for training, with early stopping applied if validation performance did not improve for 15 consecutive epochs. Data preprocessing included normalization of all EEG signals to zero mean and unit variance. To handle dataset-specific challenges, artifact removal techniques, such as Independent Component Analysis (ICA), were applied to minimize noise and enhance signal clarity. For data augmentation, methods like time-shifting, window cropping, and amplitude scaling were employed to improve generalization. Each dataset was divided into training, validation, and test sets in a 70:15:15 ratio, ensuring no overlap between participants in different splits. The model employed a batch size of 32 and a dropout rate of 0.4 for regularization. For each dataset, the input consisted of multi-channel EEG signals reshaped into fixed-length windows of 2 seconds with a sampling rate of 128 Hz. Frequency-domain features were extracted using Short-Time Fourier Transform (STFT) to complement the time-domain inputs. This multi-view input representation enhanced the model’s ability to capture both temporal and spectral information. The evaluation metrics included classification accuracy, F1-score, precision, recall, and area under the receiver operating characteristic curve (AUC) for binary and multiclass tasks. For regression tasks, mean squared error (MSE) and mean absolute error (MAE) were reported. Each metric was averaged over 5-fold cross-validation to ensure statistical robustness. The ablation study was conducted to assess the impact of individual model components on performance, as detailed in subsequent sections. Hyperparameter tuning was performed using a grid search across learning rates {0.001, 0.0005, 0.0001}, dropout rates {0.3, 0.4, 0.5}, and batch sizes {16, 32, 64}. The optimal configuration for each dataset was selected based on the best validation performance. To ensure reproducibility, all experiments were conducted with a fixed random seed of 42. The computational efficiency of the model was also evaluated by recording training and inference times. The scalability of the approach was tested on varying input lengths and channel counts, demonstrating the robustness of the architecture across different experimental setups. All experiments adhered to ethical guidelines, and the use of publicly available datasets ensured compliance with data privacy and sharing standards (Algorithm 3).
Algorithm 3. Training Process for HIDN on Multimodal EEG Datasets.
4.3 Comparison with SOTA methods
The comparative analysis of our proposed method with state-of-the-art (SOTA) approaches on the PhyAAt, Physionet MI, BIDS Siena Scalp EEG, and STEW datasets for emotion recognition is detailed in Tables 1, 2. These results demonstrate the superior performance of our approach across all metrics, including accuracy, precision, recall, and F1 score, when compared to other competitive models like LSTM, GRU, Transformer, and Informer.

Table 1. Comparison of Ours with SOTA methods on PhyAAt and Physionet MI datasets for Emotion Recognition.

Table 2. Comparison of Ours with SOTA methods on BIDS Siena Scalp EEG and STEW datasets for Emotion Recognition.
On the PhyAAt dataset, our model achieved an accuracy of 90.87%, which is significantly higher than Informer (89.01%) and Transformer (88.45%). Precision, recall, and F1 score also improved notably, with values of 89.34%, 89.76%, and 89.55%, respectively. Similarly, for the Physionet MI dataset, our approach outperformed others with an accuracy of 92.48% and an F1 score of 90.15%. These results highlight the robustness of our method in capturing intricate temporal patterns and physiological signal dependencies critical for accurate emotion recognition. For the BIDS Siena Scalp EEG dataset, our model achieved an accuracy of 90.12%, exceeding Informer (88.92%) and Transformer (87.67%). Precision and recall values further reinforce this improvement, with our model recording 88.76% and 89.34%, respectively. The STEW dataset results show our model setting a new benchmark, with an accuracy of 91.67% and an F1 score of 89.72%. This indicates its capability to generalize well across diverse datasets with varying complexities in EEG data. The enhanced performance of our method can be attributed to its novel integration of hierarchical temporal attention mechanisms and adaptive feature selection. Unlike LSTM and GRU, which struggle to capture long-range dependencies, our approach leverages multi-scale feature representation to model both short- and long-term temporal dynamics effectively. Furthermore, the incorporation of domain-specific preprocessing techniques, such as artifact removal and frequency-domain feature extraction, ensures that the model processes cleaner, more informative signals.
Figures 5, 6 visually illustrate the comparative improvements achieved by our model, particularly in terms of accuracy and F1 score. These figures emphasize the model’s ability to consistently outperform SOTA methods across diverse datasets and metrics. The results substantiate the scalability and adaptability of our proposed architecture in emotion recognition tasks, setting a new standard for EEG-based analysis in affective computing and related applications.

Figure 5. Performance comparison of SOTA methods on PhyAAt dataset and Physionet MI dataset datasets.

Figure 6. Performance comparison of SOTA methods on BIDS Siena Scalp EEG dataset and STEW dataset datasets.
4.4 Ablation study
To investigate the contributions of individual components in our model, we conducted an ablation study on the PhyAAt, Physionet MI, BIDS Siena Scalp EEG, and STEW datasets for emotion recognition. The results are summarized in Tables 3, 4, where we evaluate the performance of the complete model and its variants with one component removed at a time.
For the PhyAAt dataset, the complete model achieved an accuracy of 90.87% and an F1 score of 89.55%. When Temporal Dynamics with Recurrent Neural Structures was removed, accuracy dropped to 88.01% and the F1 score to 86.61%, indicating the critical role of Temporal Dynamics with Recurrent Neural Structures in capturing long-range dependencies and temporal attention. Similarly, removing Integration of External Stimuli through Encoding resulted in an accuracy of 89.05% and an F1 score of 87.87%, reflecting its importance in adaptive feature selection and representation. Excluding Personalization and Risk-Aware Adaptation reduced accuracy to 88.45% and the F1 score to 87.22%, highlighting its contribution to multi-scale feature integration. For the Physionet MI dataset, the complete model exhibited an accuracy of 92.48% and an F1 score of 90.15%, outperforming all ablated variants. Without Temporal Dynamics with Recurrent Neural Structures, accuracy fell to 90.02% and the F1 score to 87.63%. The exclusion of Integration of External Stimuli through Encoding led to an accuracy of 91.02% and an F1 score of 88.50%, while removing Personalization and Risk-Aware Adaptation yielded slightly lower results, with an accuracy of 90.52% and an F1 score of 88.11%. These findings underscore the synergetic effect of all components in achieving optimal performance. On the BIDS Siena Scalp EEG dataset, the complete model recorded an accuracy of 90.12% and an F1 score of 89.05%. Removing Temporal Dynamics with Recurrent Neural Structures decreased accuracy to 88.12%, while removing Components B and C resulted in accuracies of 89.01% and 88.56%, respectively. Similarly, for the STEW dataset, the complete model achieved superior performance with an accuracy of 91.67% and an F1 score of 89.72%. The removal of Temporal Dynamics with Recurrent Neural Structures reduced accuracy to 90.21%, while the absence of Components B and C led to accuracies of 91.12% and 90.78%, respectively.
The ablation study reveals that Temporal Dynamics with Recurrent Neural Structures contributes significantly to capturing long-range temporal dependencies through hierarchical attention mechanisms. Integration of External Stimuli through Encoding enhances the model’s ability to adaptively select and emphasize critical features, particularly in noisy EEG data. Personalization and Risk-Aware Adaptation enables effective multi-scale representation of temporal and spatial patterns, improving overall generalization. The full integration of these components ensures that the model captures intricate and nuanced patterns in the data, leading to state-of-the-art performance. Figures 7, 8 illustrate the performance degradation observed in ablated variants compared to the complete model, further validating the necessity of each component. This study highlights the robust and complementary design of our architecture, setting a new benchmark for emotion recognition in EEG-based systems.

Figure 7. Ablation study of our method on PhyAAt dataset and physionet MI dataset datasets.temporal dynamics with recurrent neural structures (TD), integration of external stimuli through encoding (IES), personalization and risk-aware adaptation (PRA).

Figure 8. Ablation study of our method on BIDS siena scalp EEG dataset and STEW dataset datasets. Temporal dynamics with recurrent neural structures (T), integration of external stimuli through encoding (I), personalization and risk-aware adaptation (P).
The results obtained from the experiments on the MIDUS II and PPG-DaLiA datasets demonstrate the superior performance of our proposed model in capturing hormone dynamics when compared to several state-of-the-art baselines. In Table 5, across both datasets, our method achieved the lowest RMSE and MAE values, indicating higher precision in predicting hormone concentration changes over time. On the MIDUS II dataset, our model reached an RMSE of 0.139 and MAE of 0.110, outperforming models such as Transformer and Informer, which, while competitive, exhibited higher prediction errors. A similar pattern was observed on the PPG-DaLiA dataset, where our model consistently delivered improved accuracy in forecasting cortisol fluctuations, with an RMSE of 0.128 and MAE of 0.102. In addition to numerical accuracy, our model also achieved the highest Pearson correlation coefficients (PCC) on both datasets, with values of 0.887 and 0.869, respectively. This suggests a stronger alignment between predicted and actual hormonal trends, reflecting the model’s capacity to capture both temporal dependencies and the physiological regularities embedded in endocrine patterns. The improvements observed can be attributed to the integration of domain-informed graph dynamics and temporal feedback mechanisms within the HIDN architecture, as well as the adaptability introduced by AHRS. These findings confirm that our framework is not only effective in modeling complex neuroendocrine interactions but also generalizes well across datasets with varying sampling frequencies and hormonal modalities.

Table 5. Comparison of ours with SOTA methods on MIDUS II and PPG-DaLiA datasets for hormonal dynamics prediction.
5 Discussion
The experimental results not only validate the performance of our proposed framework on emotion recognition tasks, but also offer insights into the underlying interactions between endocrine systems and EEG signals. The improved prediction accuracy observed across multiple datasets suggests that EEG features indeed encode information reflective of hormonal dynamics. Ablation studies demonstrate that modeling hormone feedback loops and incorporating external physiological inputs enhance the system’s ability to capture complex neuroendocrine dependencies. These findings support the hypothesis that emotional states—often driven by neural responses—can serve as indirect indicators of hormonal variations, and vice versa. Our work thus contributes to a deeper computational understanding of the bidirectional relationship between brain activity and endocrine regulation, opening pathways for integrated diagnostic and therapeutic strategies.
6 Conclusions and future work
This research develops a mathematical model to investigate the complex interaction between endocrine systems and EEG signals, a critical area for advancing physiological and neurological health management. Traditional models often fall short due to their inability to accurately capture nonlinear dynamics, feedback loops, and intricate cross-system interactions. To address these limitations, this study introduces a novel framework incorporating the Hormone Interaction Dynamics Network (HIDN) and the Adaptive Hormonal Regulation Strategy (AHRS). HIDN employs graph-based neural architectures and recurrent dynamics to represent spatial-temporal interdependencies among endocrine glands, hormones, and EEG signals. In parallel, AHRS enhances the framework’s adaptability through real-time feedback and patient-specific adjustments, optimizing therapeutic interventions. This dual approach significantly improves scalability, precision, and robustness, overcoming challenges such as sparse data, temporal resolution issues, and multi-hormonal complexity. Experimental results confirm the model’s effectiveness in predicting hormone dynamics, EEG patterns, and therapeutic outcomes, providing both theoretical insights and practical applications in healthcare.
However, two limitations remain. The reliance on extensive computational resources may hinder the model’s accessibility for clinical practitioners or resource-limited settings. Future efforts could focus on developing lightweight algorithms or cloud-based solutions to address this issue. The model’s reliance on sparse clinical data poses challenges in generalizability and accuracy across diverse patient populations. Expanding the dataset diversity and incorporating synthetic data generation techniques could mitigate this limitation. Addressing these challenges will strengthen the framework’s applicability, paving the way for more integrated and personalized healthcare solutions in endocrine-neurological research.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding author.
Author contributions
WL: Writing – review & editing, Visualization, Supervision, Funding acquisition.
Funding
The author(s) declare that no financial support was received for the research and/or publication of this article.
Conflict of interest
The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Tao W, Li C, Song R, Cheng J, Liu Y, Wan F, et al. Eeg-based emotion recognition via channel-wise attention and self attention. IEEE Trans Affect Computing. (2023). Available online at: https://ieeexplore.ieee.org/abstract/document/9204431/.
2. Cai Y, Li X, and Li J. Emotion recognition using different sensors, emotion models, methods and datasets: A comprehensive review. Ital Natl Conf Sensors. (2023). Available online at: https://www.mdpi.com/1424-8220/23/5/2455.
3. Li X, Zhang Y, Tiwari P, Song D, Hu B, Yang M, et al. Eeg based emotion recognition: A tutorial and review. ACM Computing Surveys. (2022). doi: 10.1145/3524499
4. Kamble KS and Sengupta J. A comprehensive survey on emotion recognition based on electroencephalograph (eeg) signals. Multimedia Tools Appl. (2023). doi: 10.1007/s11042-023-14489-9
5. Pepino L, Riera P, and Ferrer L. Emotion recognition from speech using wav2vec 2.0 embeddings. Interspeech. (2021). Available online at: https://universite-paris-saclay.hal.science/hal-04442990/.
6. Shen W, Wu S, Yang Y, and Quan X. Directed acyclic graph network for conversational emotion recognition. Annu Meeting Assoc Comput Linguistics. (2021). Available online at: https://arxiv.org/abs/2105.12907.
7. Song T, Zheng W, Song P, and Cui Z. Eeg emotion recognition using dynamical graph convolutional neural networks. IEEE Trans Affect Computing. (2020). Available online at: https://ieeexplore.ieee.org/abstract/document/8320798/.
8. Wang Z, Wang Y, Hu C, Yin Z, and Song Y. Transformers for eeg-based emotion recognition: A hierarchical spatial information learning model. IEEE Sensors J. (2022). Available online at: https://ieeexplore.ieee.org/abstract/document/9684393/.
9. Chudasama VM, Kar P, Gudmalwar A, Shah NJ, Wasnik P, and Onoe N. (2022). M2fnet: Multi-modal fusion network for emotion recognition in conversation, in: 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). Available online at: http://openaccess.thecvf.com/content/CVPR2022W/MULA/html/Chudasama_M2FNet_Multi-Modal_Fusion_Network_for_Emotion_Recognition_in_Conversation_CVPRW_2022_paper.html.
10. Zhang S, Zhao X, and Tian Q. Spontaneous speech emotion recognition using multiscale deep convolutional lstm. IEEE Trans Affect Computing. (2022). Available online at: https://ieeexplore.ieee.org/abstract/document/8873581/.
11. Issa D, Demirci M, and Yazıcı A. Speech emotion recognition with deep convolutional neural networks. Biomed Signal Process Control. (2020). Available online at: https://www.sciencedirect.com/science/article/pii/S1746809420300501.
12. Andayani F, Theng LB, Tsun MTK, and Chua C. Hybrid lstm-transformer model for emotion recognition from speech audio files. IEEE Access. (2022). Available online at: https://ieeexplore.ieee.org/abstract/document/9745599/.
13. Hu D, Wei L, and Huai X. Dialoguecrn: Contextual reasoning networks for emotion recognition in conversations. Annu Meeting Assoc Comput Linguistics. (2021). Available online at: https://arxiv.org/abs/2106.01978.
14. Dzedzickis A, Kaklauskas A, and Bučinskas V. Human emotion recognition: Review of sensors and methods. Ital Natl Conf Sensors. (2020). doi: 10.3390/s20030592
15. Zhang K, Li Y, Wang J, Cambria E, and Li X. Real-time video emotion recognition based on reinforcement learning and domain knowledge. IEEE Trans circuits Syst video Technol (Print). (2022). Available online at: https://ieeexplore.ieee.org/abstract/document/9400391/.
16. Han D, Kong Y, Han J, and Wang G. A survey of music emotion recognition. Front Comput Sci. (2022). doi: 10.1007/s11704-021-0569-4
17. Shalbafan M, Orooji M, and Kamalzadeh L. Psychosis beas a rare side effect of sildenafil: a case report. J Med Case Rep. (2022) 16:120. doi: 10.1186/s13256-022-03334-6
18. Sarkar P and Etemad A. Self-supervised ecg representation learning for emotion recognition. IEEE Trans Affect Computing. (2020). https://ieeexplore.ieee.org/abstract/document/9161416/.
19. Kosti R, Álvarez J, Recasens A, and Lapedriza Agatà. Context based emotion recognition using emotic dataset. IEEE Trans Pattern Anal Mach Intell. (2020). Available online at: https://ieeexplore.ieee.org/abstract/document/8713881/.
20. Li Y, Zheng W, Zong Y, Cui Z, Zhang T, and Zhou X. A bi-hemisphere domain adversarial neural network model for eeg emotion recognition. IEEE Trans Affect Computing. (2021). Available online at: https://ieeexplore.ieee.org/abstract/document/8567966/.
21. Shalbafan M, El Hayek S, and de Filippis R. Mental-health-related stigma and discrimination: Prevention, role, and management strategies. (2023). https://www.frontiersin.org/articles/10.3389/fpsyt.2023.1136995/full.
22. Marini M, Ansani A, Paglieri F, Caruana F, and Viola M. The impact of facemasks on emotion recognition, trust attribution and re-identification. Sci Rep. (2021). doi: 10.1038/s41598-021-84806-5
23. Liu W, Qiu J, Zheng W-L, and Lu B-L. Comparing recognition performance and robustness of multimodal deep learning models for multimodal emotion recognition. IEEE Trans Cogn Dev Syst. (2021). Available online at: https://ieeexplore.ieee.org/abstract/document/9395500/.
24. Shalbafan M, Rasoulian M, Hajebi A, Ghadirivasfi M, and Asadi S. Rethinking the psychiatry residency curriculum for community psychiatry training in Iran. Acad Psychiatry. (2024), 1–2. doi: 10.1007/s40596-024-01963-1
25. Lian Z, Liu B, and Tao J. Ctnet: Conversational transformer network for emotion recognition. IEEE/ACM Trans Audio Speech Lang Process. (2021). Available online at: https://ieeexplore.ieee.org/abstract/document/9316758/.
26. Akhand M, Roy S, Siddique N, Kamal AS, and Shimamura T. Facial emotion recognition using transfer learning in the deep cnn. Electronics. (2021). doi: 10.3390/electronics10091036
27. Pignatelli D, Carvalho BL, Palmeiro A, Barros A, Guerreiro SG, and Macut D. The complexities in genotyping of congenital adrenal hyperplasia: 21-hydroxylase deficiency. Front Endocrinol. (2019) 10:432. doi: 10.3389/fendo.2019.00432
28. Abbaschian BJ, Sierra-Sosa D, and Elmaghraby AS. Deep learning techniques for speech emotion recognition, from databases to models. Ital Natl Conf Sensors. (2021). doi: 10.3390/s21041249
29. Heinonen IH, Boushel R, and Kalliokoski KK. The circulatory and metabolic responses to hypoxia in humans–with special reference to adipose tissue physiology and obesity. Front Endocrinol. (2016) 7:116. doi: 10.3389/fendo.2016.00116
30. Wani T, Gunawan T, Qadri SAA, Kartiwi M, and Ambikairajah E. A comprehensive review of speech emotion recognition systems. IEEE Access. (2021). Available online at: https://ieeexplore.ieee.org/abstract/document/9383000/.
31. Mehendale N. Facial emotion recognition using convolutional neural networks (ferc). SN Appl Sci. (2020). doi: 10.1007/s42452-020-2234-1
32. Lv F, Chen X, Huang Y, Duan L, and Lin G. Progressive modality reinforcement for human multimodal emotion recognition from unaligned multimodal sequences. Comput Vision Pattern Recognition. (2021). Available at: http://openaccess.thecvf.com/content/CVPR2021/html/Lv_Progressive_Modality_Reinforcement_for_Human_Multimodal_Emotion_Recognition_From_Unaligned_CVPR_2021_paper.html
33. Mo L, Ma C, Wang Z, Li J, He W, Niu W, et al. Integrated bioinformatic analysis of the shared molecular mechanisms between osteoporosis and atherosclerosis. Front Endocrinol. (2022) 13:950030. doi: 10.3389/fendo.2022.950030
34. Islam M, Moni M, Islam MM, Rashed-Al-Mahfuz M, Islam MS, Hasan MK, et al. Emotion recognition from eeg signal focusing on deep learning and shallow learning techniques. IEEE Access. (2021). Available online at: https://ieeexplore.ieee.org/abstract/document/9462089/.
35. Ahuja C and Setia D. (2022). Measuring human auditory attention with eeg, in: 2022 14th International Conference on COMmunication Systems & NETworkS (COMSNETS), . pp. 774–8. IEEE. https://ieeexplore.ieee.org/abstract/document/9668363/.
36. Hammad M, Alkinani MH, Gupta BB, and Abd El-Latif AA. Myocardial infarction detection based on deep neural network on imbalanced data. Multimedia Syst. (2022) 1–13. doi: 10.1007/s00530-020-00728-8
37. Dan J, Pale U, Amirshahi A, Cappelletti W, Ingolfsson TM, Wang X, et al. Szcore: A seizure community open-source research evaluation framework for the validation of eeg-based automated seizure detection algorithms. arXiv preprint arXiv:2402.13005. (2024). Available online at: https://arxiv.org/abs/2402.13005
38. Siddhad G, Gupta A, Dogra DP, and Roy PP. Efficacy of transformer networks for classification of eeg data. Biomed Signal Process Control. (2024) 87:105488. doi: 10.1016/j.bspc.2023.105488
39. Sherstinsky A. Fundamentals of recurrent neural network (rnn) and long short-term memory (lstm) network. Physica D: Nonlinear Phenomena. (2020) 404:132306. doi: 10.1016/j.physd.2019.132306
40. Dey R and Salem FM. (2017). Gate-variants of gated recurrent unit (gru) neural networks, in: 2017 IEEE 60th international midwest symposium on circuits and systems (MWSCAS), . pp. 1597–600. IEEE. https://ieeexplore.ieee.org/abstract/document/8053243/.
41. Wang H and Hu D. (2005). Comparison of svm and ls-svm for regression, in: 2005 International conference on neural networks and brain, , Vol. 1. pp. 279–83. IEEE. https://ieeexplore.ieee.org/abstract/document/1614615/.
42. Han K, Xiao A, Wu E, Guo J, Xu C, and Wang Y. Transformer in transformer. Adv Neural Inf Process Syst. (2021) 34:15908–19. https://proceedings.neurips.cc/paper/2021/hash/854d9fca60b4bd07f9bb215d59ef5561-Abstract.html.
43. Hewage P, Behera A, Trovati M, Pereira E, Ghahremani M, Palmieri F, et al. Temporal convolutional neural (tcn) network for an effective weather forecasting using time-series data from the local weather station. Soft Computing. (2020) 24:16453–82. doi: 10.1007/s00500-020-04954-0
Keywords: endocrine systems, EEG signals, nonlinear dynamics, adaptive regulation, hormonal modeling
Citation: Liu W (2025) Mathematical modeling of the interaction between endocrine systems and EEG signals. Front. Endocrinol. 16:1543185. doi: 10.3389/fendo.2025.1543185
Received: 11 December 2024; Accepted: 21 April 2025;
Published: 04 August 2025.
Edited by:
Johannes Wolfgang Dietrich, Ruhr University Bochum, GermanyReviewed by:
Ielyaas Cloete, Universitat Autònoma de Barcelona, SpainMahmut Ozturk, Istanbul University-Cerrahpasa, Türkiye
Copyright © 2025 Liu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Wei Liu, c2NzZnl4eGhiQHNpbmEuY29t