- 1Computer Engineering Department, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute,” Kyiv, Ukraine
- 2Laboratoire Cognitions Humaine et Artificielle, Université Paris 8, Paris, France
- 3Department of Automation and Computer Science, Faculty of Electrical and Computer Engineering, Cracow University of Technology, Cracow, Poland
Analysis of electroencephalography (EEG) signals gathered by brain–computer interface (BCI) recently demonstrated that deep neural networks (DNNs) can be effectively used for investigation of time sequences for physical actions (PA) classification. In this study, the relatively simple DNN with fully connected network (FCN) components and convolutional neural network (CNN) components was considered to classify finger-palm-hand manipulations each from the grasp-and-lift (GAL) dataset. The main aim of this study was to imitate and investigate environmental influence by the proposed noise data augmentation (NDA) of two kinds: (i) natural NDA by inclusion of noise EEG data from neighboring regions by increasing the sampling size N and the different offset values for sample labeling and (ii) synthetic NDA by adding the generated Gaussian noise. The natural NDA by increasing N leads to the higher micro and macro area under the curve (AUC) for receiver operating curve values for the bigger N values than usage of synthetic NDA. The detrended fluctuation analysis (DFA) was applied to investigate the fluctuation properties and calculate the correspondent Hurst exponents H for the quantitative characterization of the fluctuation variability. H values for the low time window scales (< 2 s) are higher in comparison with ones for the bigger time window scales. For example, H more than 2–3 times higher for some PAs, i.e., it means that the shorter EEG fragments (< 2 s) demonstrate the scaling behavior of the higher complexity than the longer fragments. As far as these results were obtained by the relatively small DNN with the low resource requirements, this approach can be promising for porting such models to Edge Computing infrastructures on devices with the very limited computational resources.
1 Introduction
Recently, deep learning (DL) methods based on deep neural networks (DNNs) were effectively used for processing different data (LeCun et al., 2015). In healthcare and elderly care, they become very popular for processing the very complex multimodal medical data (Chen and Jain, 2020; Esteva et al., 2019). Usage of DL is especially important in the view of availability of various brain–computer interfaces (BCI) used for collection and analysis of electroencephalography (EEG) signals generated by brain activities (Roy et al., 2019; Kotowski et al., 2020; Lawhern et al., 2018). In the context of critically important tasks, for example, for air-space applications, BCIs are intensively used for the mental workload assessment on professional air traffic controllers during realistic air traffic control tasks (Aricò et al., 2016b,a; Di Flumeri et al., 2019).
DNNs were actively used for analysis of EEG data in a different fields (Li et al., 2020; Aggarwal and Chugh, 2022; Zabcikova et al., 2022) such as air-space (Aricò et al., 2016b,a; Di Flumeri et al., 2019), medicine (Chen et al., 2022; Wan et al., 2019; Gu et al., 2021), education (Xu and Zhong, 2018; Gang et al., 2018; Belo et al., 2021), entertainment (Kerous et al., 2018; Gang et al., 2018; Vasiljevic and de Miranda, 2020; Cattan, 2021), and other applications (Zabcikova et al., 2022). Usually, components of convolutional neural network (CNN) (Lawhern et al., 2018; Lin et al., 2020; Gu et al., 2021; Gatti et al., 2019; Gordienko et al., 2021c), recurrent neural networks (RNN) (An and Cho, 2016; Wang et al., 2018b; Pancholi et al., 2021; Kostiukevych et al., 2021), and other including components of fully connected networks (FCN) (Gordienko et al., 2021c; Kostiukevych et al., 2021) are used in them. These models combine some methods of EEG feature extraction with the use of various filters and show significant improvement of performance in comparison with other models. For instance, 3D CNN model based on multi-dimensional feature combination improves the classification accuracy of sensorimotor area activated tasks in the brain (Wei and Lin, 2020). Some of the DNN models demonstrated their quite high efficiency on some tasks such as sleep stage classification, stress recognition, fatigue detection, motor imagery classification, emotion recognition, and emotion classification (Gu et al., 2021). As to the domain operator-specific scenarios, some interesting results were obtained for EEG hand movement force and speed forecasting with the accuracy >80% (Gatti et al., 2019) and the conflict prediction accuracy ≈60% (Vahid et al., 2020).
Some hybridization approaches become popular recently. For example, CNN components combined with RNN components (including long short-term memory (LSTM) blocks) were investigated recently to resolve action classification problem. For instance, various RNN architectures were compared in performance for identifying hand motions for GAL dataset from EEG recordings (An and Cho, 2016; Kostiukevych et al., 2021) and for AJILE dataset (Wang et al., 2018b).
As it is well-known in computer vision, for example, for image classification tasks, data augmentation (DA), in general, and noise data augmentation (NDA), in part, can improve the performance of DNNs. Various strategies for applying DA methods to EEG data were considered recently that allow to improve classification accuracy when the limited volume of the data is available (George et al., 2022). NDA methods can be performed by adding Gaussian noise (Cecotti et al., 2015; Freer and Yang, 2020; Gordienko et al., 2021c) or by creating synthetic EEG data (Zhang and Liu, 2018; Aznan et al., 2019; Fahimi et al., 2020). The similar numerous NDA-related approaches were proposed (Freer and Yang, 2020; Gordienko et al., 2021c; George et al., 2022), and many others were reviewed recently in several surveys (Rommel et al., 2022; Lashgari et al., 2020; Talavera et al., 2022).
Although the results are promising and intriguing, their statistical reliability remains uncertain due to potential external influences under real-world conditions. That is why the main of this study was to imitate and investigate environmental influence by the proposed NDA of two kinds: (i) natural NDA by inclusion of noise EEG data from neighboring regions by increasing the sampling size N and the different offset values for sample labeling (see details below) and (ii) synthetic NDA by adding the generated Gaussian noise.
It should be noted that DA is a widely used technique that enhances a model's ability to generalize by making it more robust to variations in input data. Common DA methods include geometric transformations, noise-based modifications (such as roughening, adding, or mixing), and generative approaches. However, in EEG analysis, geometric transformations such as scaling, rotation, and reflection are not directly applicable. Unlike structured tables, text, or images, EEG signals are continuous and vary over time. Even after feature extraction, they remain time series data. Applying geometric transformations, such as rotation, to EEG signals would disrupt their temporal structure, compromising their meaningful features.
Among various ways for adding noise to the EEG signals in purpose of DA Li et al. (2019); Parvan et al. (2019); Ko et al. (2021); Sun and Mou (2023), the following are of great interest due to their intuitive understanding:
• inject various types of noise (such as uniform, Gaussian, Poisson, salt, and pepper noise, and various color noise types) with different parameters (for instance: mean and standard deviation).
• manipulate the time segment of interest by shifting/adding/cropping/combining operations with including/subtracting the information about background and signal.
• synthesize the signal by encoding/decoding and generative approaches.
Like geometric transformation methods, noise addition-based DA has been widely applied in successful DL studies for CV (Simonyan and Zisserman, 2014; He et al., 2016). This approach enhances DA by introducing randomly sampled noise values into the original data. Injecting structured noise patterns (e.g., white Gaussian or pink noise) with a specific signal-to-noise ratio (SNR) can alter the spectral characteristics of a time series by introducing additional frequency components to the signal spectrum (Borra et al., 2024).
In the context of DA for EEG, numerous studies were performed to investigate the impact the noise-induced DA for EEG. Some of the recent results are shortly summarized in Table 1.

Table 1. Examples of EEG classification studies (“Reference” column) with noise-based DA with various noise parameters (“Noise Type”) on some standard or custom datasets (“Dataset Reference”) with the different numbers of classes (“Nc”), neural network architectures (“NNA”), and the reported improved accuracy (“Accuracy, (%)”) by absolute values or changes (denoted with + sign).
DA by Gaussian noise involves adding Gaussian white noise to recorded EEG signals (Wang et al., 2018a). In practice, a perturbation E(t) ~ N(0, σ 2) is independently sampled for each channel and acquisition time and added to the original signal X, resulting in the augmented data: [X](t) = X(t) + E(t) Here, σ represents the standard deviation of the noise distribution. This parameter determines the magnitude of the transformation as larger values lead to greater distortion of the original signal. The primary motivation for this DA is to enhance model robustness against noise in EEG recordings, which are known to have a limited SNR. Gaussian noise augmentation is controlled by the parameter σ, which dictates the standard deviation of the sampled noise. Selecting an appropriate σ value is as crucial as choosing the DA method itself. For example, when σ exceeds 0.2, EEG signals become excessively noisy, making the DA systematically detrimental to learning (Rommel et al., 2022).
For instance, spectrogram images of motor imagery EEG have been augmented by introducing Gaussian noise (Zhang et al., 2020). White noise manifests as random fluctuations uniformly distributed across all frequencies in the EEG signal. It can originate from various sources, including thermal noise in EEG equipment, sensor artifacts, or external electrical interference. This noise reduces the SNR, making it challenging to discern neural patterns and potentially masking true neural activity. The effect is especially problematic for low-amplitude signals, such as those originating from deep brain regions.
Another method for increasing data diversity involves injecting random matrices into the raw data, typically sampled from Gaussian distributions (Okafor et al., 2017). Gaussian noise injection applies a randomly generated matrix from a Gaussian distribution to the original data as a form of DA. While these methods are straightforward and intuitive, they can sometimes exacerbate model overfitting due to the high similarity between the original and augmented data. In particular, several studies have augmented EMG signals by adding Gaussian noise to the original dataset and adjusting the SNR (Atzori et al., 2016; Zhengyi et al., 2017; Tsinganos et al., 2018).
Another study uses injecting random Gaussian noise generated based on the statistical properties of the data. The mean value of trials for the target class is computed; then, Gaussian noise with a zero mean and a standard deviation equal to the class mean (cmean in Table 1) is generated. This noise is added to randomly selected trials to create artificial frames. This simple yet effective method preserves the original waveform characteristics while introducing slight numerical variations across trials (George et al., 2022).
In other studies, DA has proven effective in addressing the challenge of limited learning caused by small training sets in EEGNet, leading to significant improvements in classification accuracy. As a result, the data were expanded by a factor of three, and the standard deviation of the added Gaussian noise was set to 0.1 (Cai et al., 2024).
The modality-wise approach was proposed in Lopez et al. (2023), where the Gaussian noise signal with zero mean is added to each sample, with its standard deviation computed modality-wise (MW in Table 1), ensuring that the augmented signal achieves a signal-to-noise ratio (SNR) of 5dB.
In addition to Gaussian noise, various types of colored noise can also be present in EEG signals due to physiological and environmental factors. These noise types typically manifest as interference, distorting the true brain activity and complicating accurate analysis and interpretation. Below are some examples of how different types of colored noise may appear in EEG signals of brain activity:
Pink noise (1/f noise) is characterized by greater power at lower frequencies, with a gradual decrease in power as the frequency increases. This type of noise can naturally arise from brain activity, particularly during resting states, or be introduced by background physiological processes such as muscle activity or skin potentials. Pink noise can dominate low-frequency bands, potentially masking slow-wave oscillations that are crucial for sleep studies or resting-state EEG analysis.
Brown (red) noise displays even more power at lower frequencies than pink noise, with a steeper decline as the frequency increases. It can arise from long-term drift in electrode potentials or baseline shifts in the EEG signal, often caused by environmental factors that affect the EEG setup. This results in large, slow oscillations that can dominate the EEG trace, potentially masking lower-frequency brain rhythms. While Gaussian (white) noise is frequently applied, the specific use of brown noise has not been widely explored in the literature, but exploring brown noise injection could potentially offer new avenues for enhancing EEG data augmentation techniques.
Blue noise is characterized by an emphasis on high frequencies, manifesting as rapid, small-amplitude fluctuations in the signal. It can originate from high-frequency environmental interference, such as electronic devices or power lines, or from muscle activity, including micro-movements of the scalp or jaw. This type of noise can mask high-frequency neural signals, such as gamma rhythms (30–100 Hz), and may lead to false-positive detections in high-frequency analyses.
Violet noise is an extreme form of high-frequency noise, with a stronger emphasis on higher frequencies than blue noise. It can be caused by electronic interference within the EEG system, such as sudden changes in electrode contact, such as detachment or movement. This noise can introduce sharp spikes or rapid fluctuations that resemble artifacts, potentially disrupting the analysis of high-frequency components, such as event-related potentials (ERPs).
In some studies, several types of noises (in addition to white Gaussian noise) were investigated (Tangermann et al., 2012; Sun et al., 2024). To increase the number of training samples and address the variability and randomness of EEG signals, several noise DA strategies were implemented (Sun et al., 2024). Specifically, the noise DA strategies were adopted to enhance EEG data by simulating various noise sources that may be encountered in real-world environments. The types of noise applied were: (1) uniform noise, (2) Gaussian noise, (3) pink noise, (4) impulse noise, and (5) power-line noise. These noise types were randomly incorporated into the processed clean EEG signals at different proportions (ranging from 10% to 70% of the average amplitude of the EEG signal), thus generating a greater number of training samples. These DA strategies not only enhance the model's robustness to existing noise in the original signals but also improve the model's generalization capabilities in the presence of unknown noises. The introduction of noise through DA strategies has a positive effect on model training, particularly with Gaussian and pink noise. This suggests that such disturbances are prevalent in real EEG data as the noise-augmented strategies enhance the diversity of the samples and the generalization ability of the model. Overall, as the intensity of the added noise increases, both pink noise and Gaussian noise initially decrease and then increase the model prediction error. The optimal results are achieved when noise is added at 30% of the average signal intensity.
In the field of DA for EEG, the use of random shifts has been explored to some extent. However, dataset shift (where the data distribution during inference differs from that during training) is common in real biosignal-based applications. To enhance robustness, probabilistic models with uncertainty quantification are adapted to assess the reliability of predictions. Despite this, evaluating the quality of the estimated uncertainty remains a challenge. Recently, the framework was proposed to assess the ability of estimated uncertainty to capture various types of biosignal dataset shifts with different magnitudes (Xia et al., 2022). Specifically, three classification tasks were used that were based on respiratory sounds and electrocardiography signals to benchmark five representative uncertainty quantification methods. Extensive experiments reveal that, while Ensemble and Bayesian models provide relatively better uncertainty estimates under dataset shifts, all the tested models fall short in offering trustworthy predictions and proper model calibration. In another study, time-axis shifts of EEG trials were applied to generate artificial signals for DA purposes (Sakai et al., 2017). Again, the effectiveness of such geometric transformations is debated. Given the non-stationary nature of EEG signals, transformations like shifting may interfere with inherent features, potentially corrupting the data (Kalashami et al., 2022). Thus, while random shifts have been used in DA for EEG, their effectiveness continues to be a subject of ongoing research and discussion.
Another approach to applying DA for EEG involves noise injection through the inclusion of neighboring regions and other manipulations with data. A recent study utilized such DA strategies to address the challenge of small sample sizes. Specifically, translations and vertical flip operations were employed to capture a broader range of temporal information. The data were extracted from 0 to 500 ms after stimulation and then translated. Five time points within the first 200 ms after stimulus onset were randomly selected, and data from 500 ms later were collected. This method increased the dataset size 6-fold. Subsequently, the data were flipped by taking the opposite value of the augmented data, further expanding the dataset to 12 times its original size (Gou et al., 2022).
Similarly, other research efforts explore different data augmentation strategies, including Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs), to generate synthetic EEG data for training (Habashi et al., 2023; Ibrahim et al., 2024).
GANs, initially introduced for image generation, have also shown promise as a potential DA solution for EEG. GANs and their variants generate artificial data by training two competing networks: a generative network and a discriminative network. The generative network takes random noise from a predefined distribution (e.g., Gaussian) and attempts to create synthetic data that resemble real samples, while the discriminative network is trained to differentiate between real and synthetic data. Through adversarial training, the generative network progressively improves, ultimately producing highly realistic EEG signals (Zhang et al., 2021; Bao et al., 2021; Carrle et al., 2023; Ibrahim et al., 2024).
VAEs offer another approach to generating synthetic EEG data. Like a conventional autoencoder, a VAE consists of an encoder that transforms raw data into a latent representation and a decoder that reconstructs the data from this latent space. To generate new samples, the VAE randomly samples points from the learned latent distribution and passes them through the decoder, which reconstructs them into novel data. Both GANs and VAEs generate new samples indirectly by learning meaningful latent representations of the original data (Bao et al., 2021; Sun and Mou, 2023).
The proposed study contributes to investigation of novel approaches for noise-based DA for EEG classification with emphasis on influence of adding the randomly generated artificial noise and the natural noise created by inclusion of neighboring EEG data segments. This exploration is crucial as it could reveal and compare how artificial and natural noise DA can impact EEG classification performance for various noise DA parameters, for example, with increased sample size, varied offsets, etc. It is especially important for lightweight DNN architectures, designed for Edge Intelligence setups, to ensure efficient EEG processing with minimal computational resources, advancing biologically relevant and computationally efficient DA methods.
For effective application of noise-based DA methods, a clear understanding of the characteristics and sources of these noise types in real-world scenarios is necessary. That is why this study is limited to the simplest noise types that can be intuitively understandable and potentially interpreted. This study aims to provide a thorough understanding of noise-based DA by Gaussian noise injection to mimic random fluctuations evenly distributed across all frequencies in the EEG signal that can be caused by the environment.
The methods of statistical analysis and detrended fluctuation analysis (DFA) are widely used to investigate the fluctuation properties of the measured metrics and calculate the correspondent Hurst exponents (Hurst, 1956). For this purpose, the relatively small DNN [that was described and analyzed in details in Gordienko et al. (2021c)] with components of FCNs and CNNs was considered to classify physical activities (namely, hand manipulations) from the grasp-and-lift (GAL) dataset (Luciw et al., 2014; Kaggle, 2020). The special attention was paid to the analysis of the previous, mid- and post-action segments of the corresponding brain activity to anticipate them before the start of the action.
Finally, this study is targeted on investigation of EEG data collected by BCI to resolve classification problem for some physical activities (namely, hand manipulations) by the relatively simple DNN. The DNN was applied for analysis of preliminary (prior-activity), current (in-activity), and following (post-activity) parts of the relevant brain EEG signals. This problem is very important in the view of complex practical conditions where EEG activity can be disturbed by other physiological activities and, especially, external environmental noise. On the one hand, such disturbances can worsen the classification performance but, on the other hand, in reverse can improve it if it will be used during training as data augmentation (DA) technique.
2 Materials and methods
In this section, several important experimental aspects are explained: the dataset with EEG brain activities for six types of physical activities, structure of the model, metrics, workflow, and data augmentation techniques.
2.1 Dataset
In this study, the open “grasp-and-lift” (GAL) dataset is used that contains information about brain activity of 12 persons (Luciw et al., 2014; Kaggle, 2020): more than 3,900 trials (monitored and measured by the sampling rate of 500 Hz) in 32 channels of the recorded EEG signals. The person tries to perform six types of physical activities, namely: “HandStart”—moves hand to an object (for example, some gadget), “FirstDigitTouch”—touches the object by finger (for example, press a button), “BothStartLoadPhase”—takes (“grasps”) the object by fingers, LiftOff—raises (“lifts”) the object by fingers, Replace—returns the object by fingers back, BothReleased–releases fingers. The data from GAL dataset were previously processed in a standard way (Kostiukevych et al., 2021; Gordienko et al., 2021c) with taking into account the correspondent time position of physical actions (actually hand movements here) and their duration (Figure 1).
It should be emphasized that these kinds of physical activity can be naively divided in three parts depending on the feasibility of their classification: the easiest (HandStart), medium (LiftOff, Replace, and BothReleased), and hardest (BothStartLoadPhase and FirstDigitTouch) classification. But BothStartLoadPhase and FirstDigitTouch activities strongly overlap in this experiment and that is why hardly can be recognized as separate activities (this is planned to be fixed by collection of the original data in the same fashion in our future research).
As a part of an explanatory data analysis (EDA), visualizing and analyzing the experimental EEG data from GAL dataset (Gramfort et al., 2013) was performed by means of MNE open-source Python (Gramfort et al., 2013). For example, all EEG data measured by the BCI-sensors with their predefined spatial position can be plotted as subtopomaps of an evoked potential trough timeline (Figure 2).

Figure 2. Topographic maps of specific time points of evoked data (from 32 EEG channels) for the considered physical actions: (A) HandStart, (B) FirstDigitTouch, (C) BothStartLoadPhase, (D) LiftOff, (E) Replace, and (F) BothReleased actions.
2.2 Models
From subtopomaps of an evoked potential trough timeline (Figure 2), one can evidently see the complex distribution of EEG sifgnals over scalp. As far as EEG signals interfere with each other due to their electromagnetic nature (Figure 2), the combinations of the data channels can be effectively used for their processing on the basis of FCN, CNN, and RNN like it was demonstrated in our previous studies (Gordienko et al., 2021c; Kostiukevych et al., 2021). In this research, the relatively small “vanilla” DNN (Gordienko et al., 2021c) was used here. The main motive for the usage of CNNs was to use convolution operations inside an EEG time sequence of each EEG channel where all 32 EEG channels were considered to be independent ones. Then, the workflows from 32 EEG channels were combined in fully connected dense layers and then transmitted to the classification layer. The idea is to use 1D convolution operations across all 32 EEG channels for each time step. The mentioned “vanilla” DNN contains three convolutional layers [with 32 filters and kernel (3,1); 64 filters and kernel (5,1); 128 filters and kernel (7,1)] followed by batch normalization and max pooling layers with pool kernel (2,1) with dropout (0.1) and FCN layers.
2.3 Metrics
Several standard metrics were used such as accuracy and loss that were calculated during validation phase of the model learning with checkpointing states for the minimal value and maximal value of loss and accuracy, respectively. In addition, the area under curve (AUC) was measured for receiver operating characteristic (ROC) with their micro and macro versions, and their mean and standard deviation values. It is important because for the given threshold, the accuracy measures the percentage of objects correctly classified, regardless of which class they belong to. As far as AUC is threshold-invariant, AUC can allow us to measure the quality of the models considered here independently from the selected classification threshold. The AUC can consider various possible thresholds and can provide the wider range of the classifier performance. During validation phase, the models with the best accuracy and loss values were saved for the testing phase. For smooth line fitting by locally weighted polynomial regression method (LOWESS) (Cleveland et al., 2017) using weighted least squares, giving more weight to points near the point whose response is being estimated and less weight to points further away.
To investigate the high level of the fluctuations of the measured metrics, that was observed in our previous studies, DFA (detrended fluctuation analysis) was applied here. DFA was proposed to study some memory effects in sequences of the complex biological structures (Peng et al., 1993). During the last decades, it was successfully used in investigation of sequences by means of the scaling properties of the fluctuation function F(n) of non-overlapping time intervals of length n. F(n) is expected to scale as nH, where H is the Hurst exponent (Hurst, 1956).
2.4 Workflow and data augmentation
The training, validation, and testing stages of the whole workflow (Figure 3) for the proposed simple DNN model were applied for the single epoch only, because the main aim was not the highest possible performance, but feasibility analysis of reliable classification under induced noise. The introduced noise was of two kinds: (i) natural NDA by inclusion of noise EEG data from neighboring regions by the different offset values (see details below) and (ii) synthetic NDA by adding the generated Gaussian noise.
During each training iteration, the callbacks were organized to save the best models (with the highest accuracy and lowest loss) for the subsequent testing stage. The number of signal samples (N) in each Input EEG time Sequence (IS) was in the range from 100 to 2,000. These ISs were collected in a random way from the whole timeline of the experimental EEG data.
To mimic the natural NDA, the labels of physical activities for each IS were defined by ground truth (GT) values in the following three locations: at the beginning, medium, and end moment inside IS. These positions were denoted by the offset values, for example, if offset = 0, then label (IS) = GT (beginning); if offset = 0.5, then label (IS) = GT (medium); if offset = 1, then label (IS) = GT (end). Actually, it allowed us to get the GT labeling without neighboring regions without the correspondent physical action (offset = 0.5), GT labeling with some neighboring regions after the correspondent physical action (offset = 0), and GT labeling with some neighboring regions before the correspondent physical action (offset = 1). Actually, GT labeling with offset = 0 and offset = 1 provides inclusion of some natural EEG noise after and before the actual physical action. Of course, EEG activity related to the physical action signal (PAS) can take place before and after the actual physical action, but the increase of the number of signal samples (N) can lead to the PAS-to-noise ratio (PAS-NR) decrease and imitate the higher influence of the natural noise.
Under these conditions, the training, validation, and testing stages were independently done in an iterative way for 20 values of N with the step of 100. Finally, N values were obtained in the range from 100 to 2,000 that resulted in 20 iterations of training, validation, and testing phases. For each instance of N, the dataset was distributed in approximate proportion of 82% (≈ 300 examples)/9% (≈ 300 examples)/9% (≈ 300 examples) for training/validation/testing sets, respectively. As a result, 20 trained models were obtained for each iteration (one model per each input sequence with N values); then, 20 sets of metrics, including AUC, and its micro and macro versions, were calculated and plots of these metrics vs. N were constructed (see below).
3 Experimental
3.1 DNN training/validation/testing stages
During EDA stage, the GAL data were preprocessed in the standard way described in details in our previous studies (Gordienko et al., 2021c; Kostiukevych et al., 2021). In Figure 2, the topographic maps of specific time points of evoked data (from 32 EEG channels) for some physical actions (FirstDigitTouch and LiftOff) are shown. Here, the most characteristic parts of EEG signals and their spatial distributions over a scalp are shown for the better understanding the very complex details of EEG brain activity.
As it was demonstrated before (Gordienko et al., 2021c; Kostiukevych et al., 2021), some physical actions (such as HandStart) are followed by very pronounced patterns with the local minimums and maximums, while many others (such as BothStartLoadPhase, LiftOff, Replace) are hardly recognizable by unique patterns. In addition, it should be noted for several actions (such as HandStart, and especially Replace and BothReleased) that significant brain activity is started some milliseconds before the correspondent movements, but it is quite dubious to make the same statement about other actions in the view of the unrecognizable different patterns. The main idea of this study is based on our previous studies (Gordienko et al., 2021c; Kostiukevych et al., 2021) and consists in the hypothesis that relatively small DNNs even can classify the EEG patterns of the currently undergoing physical actions in the presence of some induced noise even, but the additional aspects include the investigation of impact induced by natural and synthetic kinds of noise.
At testing stage, AUC values were measured (dotted lines in Figure 4) and their smoothed fits were obtained by LOWESS-method (solid lines in Figure 4). For various actions and offsets, AUC values (Figure 4) demonstrate the high intensity of fluctuations with increase of N that can be explained by the influence of the non-relevant (to the current physical activity) regions of the increased time sequence under investigation (imitating the natural noise addition).

Figure 4. Comparison of AUC values as a function of the number of samples for the different physical actions (dotted lines) and their smoothed fits by LOWESS-method (solid lines) for the offsets imitating the natural noise addition: (A) 0, (B) 0.5, and (C) 1.
3.2 Noise data augmentation
The effect of the natural noise addition (by offsets 0 and 1 with various levels by increasing N) can be observed by calculation of the correspondent macro AUC (Figure 5A) and micro AUC (Figure 5B) values. It is evident that offsets 0 and 1 lead to the lower micro and macro AUC values in comparison with the GT labeling by offset = 0.5, and the AUC decrease is higher for the higher N values.

Figure 5. Comparison of macro (A) and micro (B) AUC values as a function of the number of samples for the different offset values (colors) without the synthetic noise (dotted lines) and with the synthetic noise σsynth = 0.2 (solid lines).
To investigate stability of the results obtained, the additional artificial “synthetic” NDA as the values generated with mean = 0 and different standard deviations σsynth (such as 0.001, 0.01, 0.1, and 0.2) was applied to the original normalized data (Figure 6).

Figure 6. Comparison AUC values as a function of the number of samples for the different offset values (colors) without the synthetic noise (A) and with the synthetic noise σsynth = 0.2 (B).
Calculation of maximal (Figure 7), mean (Figure 8), minimal, and range (Figure 9), and standard deviation (Figure 10) of AUC values was performed for the different offset values and synthetic noise σsynth values.

Figure 7. Comparison of maximal AUC values as a function of the number of samples for the different offset values (colors) without the synthetic noise (A) and with the synthetic noise σsynth = 0.2 (B). The error bars denote the standard deviation values.

Figure 8. Comparison of mean AUC values as a function of the number of samples for the different offset values (colors) without the synthetic noise (A) and with the synthetic noise σsynth = 0.2 (B). The error bars denote the standard deviation values.

Figure 9. Comparison of range AUC values as a function of the number of samples for the different offset values (colors) without the synthetic noise (A) and with the synthetic noise σsynth = 0.2 (B). The error bars denote the standard deviation values.

Figure 10. Comparison of standard deviation AUC values as a function of the number of samples for the different offset values (colors) without the synthetic noise (A) and with the synthetic noise σsynth = 0.2 (B).
From the results obtained, the similar tendency can be observed for all actions: maximal AUC values (Figure 7) are higher for offset = 0.5 than for other offset values (0 and 1) and significantly bigger than standard deviation values. The maximal AUC values for offset = 0 are slightly higher than for offset = 1, but these differences cannot be considered as statistically significant and they are in the limits of standard deviation values.
The mean AUC values (Figure 8) are even more higher for offset = 0.5 than for other offset values (0 and 1) and significantly bigger than standard deviation values also. The mean AUC values for offset = 0 are even more higher than for offset = 1, but again these differences cannot be considered as statistically significant and they are in the limits of standard deviation values.
On the contrary, the range AUC values (Figure 9), which are differences between maximal and minimal AUC values, are lower for offset = 0.5 than for other offset values (0 and 1), and these differences are significantly bigger than standard deviation values. Similarly, the range AUC values for offset = 0 are lower than for offset = 1, and these differences are also statistically significant and beyond the limits of standard deviation values.
As it was seen from the previous Figure 9, the standard deviation AUC values (Figure 9) are significantly lower for offset = 0.5 than for other offset values (0 and 1). Similarly, the standard deviation AUC values for offset = 0 are significantly lower than for offset = 1.
To analyze the metrics for the steady region for N in the range from 1,000 to 10,000 samples, AUC (mean ± stdev) values were calculated (Table 2) along with the other metrics such as maximal AUC values and ranges (differences between maximal and minimal AUC values) (Table 3). The bold font in Table 3 denotes the highest values for the same level of σsynth, and the italic font does the lowest ones. HandStart action demonstrates the highest AUC values, and FirstDigitTouch and BothStartLoadPhase demonstrates the lowest ones. The important aspect is that for different actions, implication of the natural noise (presented by offsets) leads to the different consequences. For example, Replace and BothReleased actions have the lowest AUC decrease and highest AUC values for offset 0. As a possible explanation for these results, Replace and BothReleased actions can have the higher PAS-NR (in comparison with other actions) for offset 0 because of the more EEG activity for “after” (post-process) part of the relevant sampling. At the same time, mediocre performance for FirstDigitTouch and BothStartLoadPhase can be explained by their coincidence in time (see Figure 1) that is the real drawback of the GAL dataset used.

Table 2. AUC (mean ± stdev) (Figures 8, 10) values for the steady region (Figure 4) from 1,000 to 10,000 samples.

Table 3. Maximal and range AUC values for the steady region from 1,000 to 10,000 samples (Figures 4, 11, 12).
3.3 Detrended fluctuation analysis
Detrended Fluctuation Analysis (DFA) (Peng et al., 1993, 1995; Bianchi, 2020) was applied to analyze fluctuations of AUC values and the correspondent Hurst exponent values after eliminating the temporal trend (Figure 13). For the time sequences, the Hurst exponent value (H) can indicate whether a process is persistent or anti-persistent, but here the Hurst exponent is used for the other purpose, namely, for quantitative estimation of fluctuations svariability.
In general, the Hurst exponent, H, is intrinsically related to the fractal dimension, which quantifies the “roughness” or variability of a time series (Hurst, 1956). Specifically, the value of H provides insight into the degree of smoothness in the data: Sequences that exhibit greater variability and are more irregular (i.e., more jagged) are associated with lower values of H, approaching zero. Conversely, smoother sequences yield values of H closer to one. This relationship between H and the fractal dimension is instrumental in characterizing the long-term dependence and self-similarity in stochastic processes. The Hurst exponent can also characterize a process (Bianchi, 2020) depending on the range of the measured values: H in the range [0.0, 0.5) corresponds to a very noisy process; the value H = 0.5 relates to uncorrelated process; H in the range (0.5, 1.0] relates to persistency where long-range correlations and relatively little noise can be observed; and H > 1.0 characterizes a non-stationary process with stronger long-range correlations. The correspondent open-source Python package “fathon” was used for DFA and further analysis of metric fluctuations (Bianchi, 2020).
4 Discussion
The results obtained show that different actions can be classified with the quite different reliability. The different kinds of physical activity take the different level of physical activation and the correspondent EEG activity, for example, HandStart (fingers, palm, forearm, and shoulder are activated) includes involvement of more limbs than LiftOff (fingers, palm, and forearm) and Replace (fingers, palm, and forearm), and even more than BothReleased (several fingers and palm), BothStartLoadPhase (two fingers), and FirstDigitTouch (one finger). It should be noted that the observed performance of classification demonstrates some correlation where the higher performance by AUC (Figure 4) corresponds to the more pronounced physical activity in the following order from the highest AUC values to the lowest ones: HandStart → LiftOff → Replace → BothReleased → BothStartLoadPhase → FirstDigitTouch (Figure 4).
In addition, for N values in the range [100, 1,500], HandStart action demonstrates the asymmetric behavior with regard to the offset values 0 and 1 (Figure 4C), namely: AUC values grows much faster with N for offset = 1 (the dotted and smoothed red lines in Figure 4A) than for offset = 0 (the dotted and smoothed red lines in Figure 4C). It means that the related brain activity measured as “before” (pre-process) part of the correspondent EEG time sequences is more pronounced than “after” (post-process) part. As a result, this phenomenon allows us to classify HandStart before the actual physical action even as it was assumed in our previous studies (Gordienko et al., 2021c,b; Kostiukevych et al., 2021). It is in contrary to the kinds of activities that demonstrate similar behavior: similar growth of AUC values for N values in the range [100, 1,500] the offset values 0 and 1, and decay for N > 1500.
In general, AUC values are higher for the offset value 0.5 (in comparison with the offset values 0 and 1), steadily for N values in the range [100, 1,500] and nearly constant for N > 1, 500 for all kinds of activities (Figure 4B) as it was also shown in our previous studies (Gordienko et al., 2021c; Kostiukevych et al., 2021). That is why labeling by the offset of 0.5 seems to be the more significant for the classification problem. It should be noted that the uncertainty of AUC values estimated as their standard deviations decreases with an increase of N up to N = 2, 000 for offset = 0.5 and up to N = 3, 000 for offset = 0 and offset = 1. It should be noted the visually very pronounced fluctuations of all these metrics with N can be explained by the influence of the non-relevant (to the current physical activity) regions of the increased time sequence under investigation.
It should be noted that application of the natural NDA by increasing N leads to the better micro and macro AUC values for N values beyond the physical action duration which is ~ 0.3 s (that is equal to ~ N=150 samples, see Figure 1) and up to ~ 3 s (N = 1,500) for offset = 0.5. For example, micro and macro AUC values are equal to ~ 0.65 for sampling length N = 200 samples (that corresponds to ~ 0.4 s), and increase of N up to N = 1,500 leads to the better micro and macro AUC values equal to ~ 0.87 (Figures 5, 6). But to the moment it is unclear whether this improvement caused by the availability of EEG signals relevant to the physical action beyond action itself or by natural NDA. The additional interesting aspect is that micro and macro AUC values are much lower for the offset 0 and 1 (in comparison with offset = 0.5), but AUC values are improving with N (Figures 5, 6) up to ~ 6–7 s (N = 3,000) for offset = 0 and offset = 1. It means that heavy bias of labeling is not useful because it leads to distortion of PAS-NR due to the lower signal and higher noise values. Application of the synthetic NDA (Figures 5, 6) in the wide range of noise levels (σsynth from 0.001 up to 0.2) demonstrates the general stability of the DNN used for classification of all activities with the similar micro and macro AUC values in the limits of their fluctuations.
AUC fluctuations caused by the added synthetic NDA, shown in Figures 11, 12, are not significant in comparison with AUC fluctuations without synthetic NDA due to increase of sampling size N.

Figure 11. Noise influence on AUC values for offset = 0.5: (A) HandStart, (B) FirstDigitTouch, and (C) BothStartLoadPhase.

Figure 12. Noise influence on AUC values for offset=0.5 (continued from Figure 11): (A) LiftOff, (B) replace, (C) BothReleased.
To characterize AUC fluctuations (Figures 11, 12) with regard to the added synthetic NDA, the DFA was applied and analyzed for original (non-added noise) EEG time sequences (Figure 13A) and ones with NDA (Figure 13B). From DFA point of view, some very intensive actions (such as HandStart and LiftOff) demonstrate the very high stability to noise data augmentation with negligible changes of fluctuation amplitudes measured like differences (Figure 13) between the correspondent AUC fluctuation values for original (without added noise) (Figure 13A) and noise-augmented EEG time sequences (Figure 13B).

Figure 13. Fluctuations vs. the number of samples (N) in the input for various actions and levels (standard deviations) of the synthetic noise: (A) σsynth = 0 (no noise), (B) σsynth = 0.2, (C) difference of fluctuations from previous regimes without and with noise. The legends contain Hurst exponent values.
For FirstDigitTouch, BothStartLoadPhase, and Replace activities, the synthetic NDA actually lead to decrease of the AUC fluctuations (Figure 13C) with slow decay of this improvement with increase of N (due to above mentioned non-relevant noisy neighboring regions, i.e., the natural NDA). In contrary, for HandStart and BothReleased activities, the synthetic NDA actually lead to increase of the AUC fluctuations (Figure 13C) with slow decay also. Liftoff activity does not demonstrate any significant changes.
The measurements of the Hurst exponent values Hfull (the full range of window scales with n < 10, 000, time < 20 s), Hlow (the low window scales with n < 1, 000, time < 2 s), and Hhigh (the bigger window scales with n > 1, 000, time > 2 s) were performed for various actions and levels (standard deviations σsynth = 0) of the synthetic noise (Table 4). H values are rounded to 2 decimal digits in Table 4 because the bigger number of significant digits seems to be statistically insignificant.

Table 4. Hurst exponents Hfull, Hlow, and Hhigh (rounded to 2 decimal digits) for various actions and levels (by standard deviations σsynth) of the synthetic noise.
The general tendency is that for the low window scales (n < 1, 000, time < 2 s), Hlow values are higher in comparison with Hhigh values for the bigger window scales (n > 1, 000, time > 2 s) that can be seen from the slope of curves in Figure 13 and Table 4. It means that EEG fragments with the duration of scale n < 1, 000 (time < 2 s) demonstrate the scaling behavior of the higher complexity than the fragments n > 1, 000 (time > 2 s), i.e., Hlow > Hhigh (Table 4). It should be noted that step-like increases in the middle and in the end of all curves in Figure 13 can be explained by overlapping with the next portion of PAS data related to the other trial of recorded physical activities which are contained in the whole timeline of the experimental EEG data.
Despite the previously mentioned findings, the study has several limitations that should be taken into account in future works. First, the use of a single epoch for training (which was observed to be enough for saturation of the training process of the relatively small DNN with the quite small capacity) may limit the model's overall performance. In future research, the multiple training epochs for the more complex DNN should be employed to potentially improve model performance and generalization with attention to the impact of hyperparameter tuning (e.g., learning rate, batch size) on model performance and convergence. Second, the focus on feasibility analysis rather than maximizing performance might have constrained the exploration of more complex DNN architectures. In the next stage of the investigation, the more complex DNN architectures (such as deeper CNNs, recurrent neural networks, and transformer models) should be performed with a more comprehensive hyperparameter search to optimize model performance to potentially achieve higher classification accuracy. Third, the study relies mainly on the GAL dataset, which may not fully capture the variability and complexity of real–world EEG signals. In the extended version, this study should include investigation of the model's performance on other publicly available EEG datasets with different characteristics in the other controlled and realistic environment to improve the generalizability of the findings. Fourth, the analysis is limited to a specific set of NDA techniques, and the impact of other noise sources or more sophisticated DA methods should be also explored. Moreover, the impact of other NDA techniques (mentioned in the introductory part of the study, such as generative training and others) will be necessary to improve model robustness and explore the impact of physiological noise (e.g., muscle artifacts, eye blinks) and environmental noise on model performance. In addition, assessing the additional metrics particularly with regard to Structural Similarity Index (SSIM) and Peak Signal-to-Noise Ratio (PSNR) in future research stages will be highly intriguing and valuable. SSIM could provide insights into the structural similarity between original and noise-augmented EEG signals, helping evaluate how natural noise preserves critical signal features. Meanwhile, PSNR could serve as a measure of distortion, indicating how much the augmented signals deviate from the original ones, which is crucial for maintaining signal integrity in classification tasks.
5 Conclusion
This research contributes to the field of EEG-based BCI by investigating the impact of different types of noise on the classification of physical activities by the following main novel aspects and contributions: systematic investigation of natural noise, quantitative analysis of noise impact, and analysis of offset effects. The study introduces the concept of “natural noise” by considering EEG data from neighboring regions, simulating real-world scenarios with varying levels of background EEG activity. The researchers utilize metrics such as AUC and DFA to quantitatively assess the impact of both natural and synthetic noise on classification performance, providing valuable insights into the model's robustness. By analyzing the impact of different label offsets (0, 0.5, 1), the study provides valuable insights into the optimal time window for EEG signal analysis and classification. These novel aspects contribute to a better understanding of the challenges and limitations of EEG-based BCI systems in real-world scenarios and provide valuable guidance for future research in this area.
The following key aspects of the methodology contribute to achieving the goal: DA by natural NDA and synthetic NDA, varying NDA parameters including input sequence length, and thorough performance evaluation including DFA Analysis. The introduction of both natural and synthetic noise during DA helps the model to become more robust and generalize better to real-world scenarios with varying levels of noise. As to the natural NDA by including EEG data from neighboring regions, the model learns to handle variations in EEG signals due to temporal shifts and contextual influences. As to the synthetic NDA, adding Gaussian noise increases the model's tolerance to random fluctuations and noise in the EEG data. The use of input sequences with varying lengths (N) allows the model to assess its performance under different levels of “natural noise” introduced by the inclusion of irrelevant EEG data. This helps to understand how the model's performance is affected by the amount of surrounding EEG data. For performance evaluation, the use of multiple metrics, including AUC (micro and macro), accuracy, and loss, provides a comprehensive evaluation of the model's performance. For DFA, analysis helps to quantify the variability and complexity of the AUC fluctuations, providing insights into the model's behavior under different noise conditions. By incorporating these techniques, the authors aim to understand the feasibility and limitations of classifying EEG signals related to physical activities in the presence of noise, which is crucial for the practical application of BCI systems in real-world settings.
The results obtained allow us to conclude that the relatively simple DNN with components of FCN and CNN even can be effectively used to classify physical activities (namely, hand manipulations) from the GAL dataset. Application of natural and synthetic noises imitates the possible influence from environment. It should be noted that synthetic noise influence (due to Gaussian NDA with higher σ values) has the lower impact on the general ability to provide the better reliable classification of physical activities than natural noise influence (due to increase of the sampling size N) that can significantly improve the performance with reaching the stable metric values after some noise increase.
AUC fluctuations caused by the added synthetic NDA are not significant in comparison with AUC fluctuations without synthetic NDA due to increase of sampling size N. It should be emphasized that application of the natural NDA by increasing N leads to the better micro and macro AUC values for N values beyond the action duration which is ~ 0.3 s and up to ~ 3 s (N = 1,500) for offset = 0.5. But to the moment the open question is whether this improvement caused by the availability of EEG signals relevant to the physical action beyond action itself or by natural NDA. This aspect should be resolved by further investigations and on other open EEG datasets.
Application of the synthetic NDA in the wide range of noise levels (σsynth from 0.001 up to 0.2) demonstrates the general stability of the DNN used for classification of all activities with the similar micro and macro AUC values in the limits of their fluctuations.
DFA allows us to investigate the fluctuation properties and calculate the correspondent Hurst exponents for the quantitative characterization of their variability. As a result of this research, some PAs can be divided in separate groups of actions that can be characterized by complexity and the feasibility of their classification: the easiest (HandStart), medium (LiftOff, Replace, and BothReleased), and hardest (BothStartLoadPhase and FirstDigitTouch) classification.
A general trend is observed in the behavior of the Hurst exponent H across varying time window scales in EEG data. Specifically, for shorter time window scales (i.e., < 2 s), the values of Hlow tend to be significantly higher than those for longer time window scales (i.e., > 2 s), denoted as Hhigh. This suggests that EEG segments with durations shorter than 2 s exhibit greater scaling complexity than those of longer durations. In particular, Hlow can exceed Hhigh by a factor of 2 to 3 during certain physical actions, indicating a marked increase in complexity for these shorter time-scale fragments.
In general, this approach of adding natural noise by extending sampling size for small DNNs can be used during porting such models to Edge Computing infrastructures on devices with the very limited computational resources because the statistically reliable results were obtained by the relatively small DNN with the low resource requirements (Kochura et al., 2019; Gordienko et al., 2020, 2021a). The additional possible improvement can be obtained due to analysis of the optimal configuration for training and inference stages of the whole workflow that is especially important for distributed infrastructures (Kochura et al., 2017b; Taran et al., 2017; Gordienko et al., 2021a; Kochura et al., 2017a). Similar research could be also useful for classification of GAL-like and any other PAs before their actual start when some prediction with PA classification can be performed on the EEG activity before PA even. By this approach, the human EEG activity can be estimated with some proactive feedback such as continuation of PAs which were initiated by brain only, but unfortunately the PAs were not continued due to fatigue or some limited physical abilities, but the future detailed investigation should be performed to take into account the more various kinds of PAs.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.
Author contributions
YG: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing. NG: Investigation, Methodology, Software, Visualization, Writing – original draft, Writing – review & editing. VT: Investigation, Software, Writing – original draft, Writing – review & editing. AR: Writing – original draft, Writing – review & editing. ST: Conceptualization, Funding acquisition, Project administration, Writing – original draft, Writing – review & editing. SS: Conceptualization, Funding acquisition, Investigation, Methodology, Resources, Supervision, Validation, Writing – original draft, Writing – review & editing.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. The work was partially supported by the Knowledge At the Tip of Your Fingers: Clinical Knowledge for Humanity (KATY) project funded by the European Union's Horizon 2020 research and innovation program under grant agreement No. 101017453 in the part of research on the new neural network architectures and by the Development of hybrid models of artificial intelligence for the analysis of multimodal medical data project (K-I-144) funded by the Ministry of Education and Science of Ukraine in the part of multimodality research for medical applications.
Acknowledgments
The authors extend their appreciation to Cracow University of Technology (Cracow, Poland) and Université Paris 8 (Paris, France) that aided the efforts of the authors.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Gen AI was used in the creation of this manuscript.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Aggarwal, S., and Chugh, N. (2022). Review of machine learning techniques for EEG based brain computer interface. Arch. Comput. Methods Eng. 29, 3001–3020. doi: 10.1007/s11831-021-09684-6
An, J., and Cho, S. (2016). “Hand motion identification of grasp-and-lift task from electroencephalography recordings using recurrent neural networks,” in 2016 International Conference on Big Data and Smart Computing (BigComp) (IEEE), 427–429. doi: 10.1109/BIGCOMP.2016.7425963
Andrzejak, R. G., Lehnertz, K., Mormann, F., Rieke, C., David, P., and Elger, C. E. (2001). Indications of nonlinear deterministic and finite-dimensional structures in time series of brain electrical activity: dependence on recording region and brain state. Phys. Rev. E 64:061907. doi: 10.1103/PhysRevE.64.061907
Aricó, P., Borghini, G., Di Flumeri, G., Colosimo, A., Bonelli, S., Golfetti, A., et al. (2016a). Adaptive automation triggered by EEG-based mental workload index: a passive brain-computer interface application in realistic air traffic control environment. Front. Hum. Neurosci. 10:539. doi: 10.3389/fnhum.2016.00539
Aricó, P., Borghini, G., Di Flumeri, G., Colosimo, A., Pozzi, S., and Babiloni, F. (2016b). A passive brain-computer interface application for the mental workload assessment on professional air traffic controllers during realistic air traffic control tasks. Prog. Brain Res. 228, 295–328. doi: 10.1016/bs.pbr.2016.04.021
Ashfaq, N., Khan, M. H., and Nisar, M. A. (2024). Identification of optimal data augmentation techniques for multimodal time-series sensory data: a framework. Information 15:343. doi: 10.3390/info15060343
Atzori, M., Cognolato, M., and Müller, H. (2016). Deep learning with convolutional neural networks applied to electromyography data: A resource for the classification of movements for prosthetic hands. Front. Neurorobot. 10:9. doi: 10.3389/fnbot.2016.00009
Aznan, N. K. N., Atapour-Abarghouei, A., Bonner, S., Connolly, J. D., Al Moubayed, N., and Breckon, T. P. (2019). “Simulating brain signals: creating synthetic EEG data via neural-based generative models for improved ssvep classification,” in 2019 International Joint Conference on Neural Networks (IJCNN) (IEEE), 1–8. doi: 10.1109/IJCNN.2019.8852227
Bao, G., Yan, B., Tong, L., Shu, J., Wang, L., Yang, K., et al. (2021). Data augmentation for EEG-based emotion recognition using generative adversarial networks. Front. Comput. Neurosci. 15:723843. doi: 10.3389/fncom.2021.723843
Behncke, J., Schirrmeister, R. T., Burgard, W., and Ball, T. (2018). “The signature of robot action success in EEG signals of a human observer: decoding and visualization using deep convolutional neural networks,” in 2018 6th International Conference on Brain-Computer Interface (BCI) (IEEE), 1–6. doi: 10.1109/IWW-BCI.2018.8311531
Belo, J., Clerc, M., and Schön, D. (2021). EEG-based auditory attention detection and its possible future applications for passive BCI. Front. Comput. Sci. 3:661178. doi: 10.3389/fcomp.2021.661178
Bianchi, S. (2020). FATHON: a python package for a fast computation of detrendend fluctuation analysis and related algorithms. J. Open Source Softw. 5:1828. doi: 10.21105/joss.01828
Borra, D., Paissan, F., and Ravanelli, M. (2024). Speechbrain-moabb: An open-source python library for benchmarking deep neural networks applied to EEG signals. Comput. Biol. Med. 182:109097. doi: 10.1016/j.compbiomed.2024.109097
Brunner, C., Leeb, R., Müller-Putz, G., Schlögl, A., and Pfurtscheller, G. (2008). BCI competition 2008-graz data set a. Institute for knowledge discovery (laboratory of brain-computer interfaces), Graz University of Technology, 1–6.
Cai, Q., Liu, C., and Chen, A. (2024). Classification of motor imagery tasks derived from unilateral upper limb based on a weight-optimized learning model. J. Integr. Neurosci. 23:106. doi: 10.31083/j.jin2305106
Carrle, F. P., Hollenbenders, Y., and Reichenbach, A. (2023). Generation of synthetic EEG data for training algorithms supporting the diagnosis of major depressive disorder. Front. Neurosci. 17:1219133. doi: 10.3389/fnins.2023.1219133
Cattan, G. (2021). The use of brain-computer interfaces in games is not ready for the general public. Front. Comput. Sci. 3:628773. doi: 10.3389/fcomp.2021.628773
Cecotti, H., Marathe, A. R., and Ries, A. J. (2015). Optimization of single-trial detection of event-related potentials through artificial trials. IEEE Trans. Biomed. Eng. 62, 2170–2176. doi: 10.1109/TBME.2015.2417054
Chen, X., Li, C., Liu, A., McKeown, M. J., Qian, R., and Wang, Z. J. (2022). Toward open-world electroencephalogram decoding via deep learning: a comprehensive survey. IEEE Signal Process. Mag. 39, 117–134. doi: 10.1109/MSP.2021.3134629
Chen, Y.-W., and Jain, L. C. (2020). Deep Learning in Healthcare. Cham: Springer. doi: 10.1007/978-3-030-32606-7
Cho, H., Ahn, M., Ahn, S., Kwon, M., and Jun, S. (2017a). Supporting data for “EEG datasets for motor imagery brain computer interface.” GigaScience Datab. 10:100295.
Cho, H., Ahn, M., Ahn, S., Kwon, M., and Jun, S. C. (2017b). EEG datasets for motor imagery brain-computer interface. GigaScience 6:gix034. doi: 10.1093/gigascience/gix034
Cleveland, W. S., Grosse, E., and Shyu, W. M. (2017). “Local regression models,” in Statistical Models in S (Routledge), 309–376. doi: 10.1201/9780203738535-8
Collazos-Huertas, D. F., Álvarez-Meza, A. M., Cárdenas-Peña, D. A., Castaño-Duque, G. A., and Castellanos-Domínguez, C. G. (2023). Posthoc interpretability of neural responses by grouping subject motor imagery skills using cnn-based connectivity. Sensors 23:2750. doi: 10.3390/s23052750
Di Flumeri, G., De Crescenzio, F., Berberian, B., Ohneiser, O., Kramer, J., Aricò, P., et al. (2019). Brain-computer interface-based adaptive automation to prevent out-of-the-loop phenomenon in air traffic controllers dealing with highly automated systems. Front. Hum. Neurosci. 13:296. doi: 10.3389/fnhum.2019.00296
Esteva, A., Robicquet, A., Ramsundar, B., Kuleshov, V., DePristo, M., Chou, K., et al. (2019). A guide to deep learning in healthcare. Nat. Med. 25, 24–29. doi: 10.1038/s41591-018-0316-z
Fahimi, F., Dosen, S., Ang, K. K., Mrachacz-Kersting, N., and Guan, C. (2020). Generative adversarial networks-based data augmentation for brain-computer interface. IEEE Trans. Neural Netw. Learn. Syst. 32, 4039–4051. doi: 10.1109/TNNLS.2020.3016666
Falaschetti, L., Biagetti, G., Alessandrini, M., Turchetti, C., Luzzi, S., and Crippa, P. (2024). Multi-class detection of neurodegenerative diseases from EEG signals using lightweight lstm neural networks. Sensors 24:6721. doi: 10.3390/s24206721
Freer, D., and Yang, G.-Z. (2020). Data augmentation for self-paced motor imagery classification with c-lstm. J. Neural Eng. 17:016041. doi: 10.1088/1741-2552/ab57c0
Gang, P., Hui, J., Stirenko, S., Gordienko, Y., Shemsedinov, T., Alienin, O., et al. (2018). “User-driven intelligent interface on the basis of multimodal augmented reality and brain-computer interaction for people with functional disabilities,” in Future of Information and Communication Conference (Cham: Springer), 612–631. doi: 10.1007/978-3-030-03402-3_43
Gatti, R., Atum, Y., Schiaffino, L., Jochumsen, M., and Manresa, J. B. (2019). Prediction of hand movement speed and force from single-trial EEG with convolutional neural networks. bioRxiv, 492660. doi: 10.1101/492660
George, O., Smith, R., Madiraju, P., Yahyasoltani, N., and Ahamed, S. I. (2022). Data augmentation strategies for EEG-based motor imagery decoding. Heliyon 8:e10240. doi: 10.1016/j.heliyon.2022.e10240
Gordienko, Y., Kochura, Y., Taran, V., Gordienko, N., Rokovyi, A., Alienin, O., et al. (2020). “Scaling analysis of specialized tensor processing architectures for deep learning models,” in Deep Learning: Concepts and Architectures (Cham: Springer), 65–99. doi: 10.1007/978-3-030-31756-0_3
Gordienko, Y., Kochura, Y., Taran, V., Gordienko, N., Rokovyi, O., Alienin, O., et al. (2021a). “last mile” optimization of edge computing ecosystem with deep learning models and specialized tensor processing architectures. Adv. Comput. 122, 303–341. doi: 10.1016/bs.adcom.2020.10.003
Gordienko, Y., Kostiukevych, K., Gordienko, N., Rokovyi, O., Alienin, O., and Stirenko, S. (2021b). “Deep learning for grasp-and-lift movement forecasting based on electroencephalography by brain-computer interface,” in International Conference on Artificial Intelligence and Logistics Engineering (Cham: Springer), 3–12. doi: 10.1007/978-3-030-80475-6_1
Gordienko, Y., Kostiukevych, K., Gordienko, N., Rokovyi, O., Alienin, O., and Stirenko, S. (2021c). “Deep learning with noise data augmentation and detrended fluctuation analysis for physical action classification by brain-computer interface,” in 2021 8th International Conference on Soft Computing &Machine Intelligence (ISCMI) (IEEE), 176–180. doi: 10.1109/ISCMI53840.2021.9654829
Gou, H., Piao, Y., Ren, J., Zhao, Q., Chen, Y., Liu, C., et al. (2022). A solution to supervised motor imagery task in the bci controlled robot contest in world robot contest. Brain Sci. Adv. 8, 153–161. doi: 10.26599/BSA.2022.9050014
Gramfort, A., Luessi, M., Larson, E., Engemann, D. A., Strohmeier, D., Brodbeck, C., et al. (2013). MEG and EEG data analysis with Mne-python. Front. Neurosci. 7:267. doi: 10.3389/fnins.2013.00267
Gu, X., Cao, Z., Jolfaei, A., Xu, P., Wu, D., Jung, T.-P., et al. (2021). EEG-based brain-computer interfaces (BCIS): a survey of recent studies on signal sensing technologies and computational intelligence approaches and their applications. IEEE/ACM Trans. Comput. Biol. Bioinform. 18, 1645–1666. doi: 10.1109/TCBB.2021.3052811
Habashi, A. G., Azab, A. M., Eldawlatly, S., and Aly, G. M. (2023). Generative adversarial networks in EEG analysis: an overview. J. Neuroeng. Rehabil. 20:40. doi: 10.1186/s12984-023-01169-w
He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on Computer Vision and Pattern Recognition 770–778. doi: 10.1109/CVPR.2016.90
Hurst, H. E. (1956). The problem of long-term storage in reservoirs. Hydrol. Sci. J. 1, 13–27. doi: 10.1080/02626665609493644
Ibrahim, M., Khalil, Y. A., Amirrajab, S., Sun, C., Breeuwer, M., Pluim, J., et al. (2024). Generative ai for synthetic data across multiple medical modalities: a systematic review of recent developments and challenges. arXiv preprint arXiv:2407.00116.
Jayaram, V., and Barachant, A. (2018). Moabb: trustworthy algorithm benchmarking for bcis. J. Neural Eng. 15:066011. doi: 10.1088/1741-2552/aadea0
Jeong, J.-H., Cho, J.-H., Lee, Y.-E., Lee, S.-H., Shin, G.-H., Kweon, Y.-S., et al. (2022). 2020 international brain-computer interface competition: a review. Front. Hum. Neurosci. 16:898300. doi: 10.3389/fnhum.2022.898300
Kaggle (2020). Grasp-and-lift EEG detection. Available at: https://www.kaggle.com/competitions/grasp-and-lift-eeg-detection (accessed October 14, 2020).
Kalashami, M. P., Pedram, M. M., and Sadr, H. (2022). EEG feature extraction and data augmentation in emotion recognition. Comput. Intell. Neurosci. 2022:7028517. doi: 10.1155/2022/7028517
Kaya, M., Binli, M. K., Ozbay, E., Yanar, H., and Mishchenko, Y. (2018). A large electroencephalographic motor imagery dataset for electroencephalographic brain computer interfaces. Sci. Data 5, 1–16. doi: 10.1038/sdata.2018.211
Kemp, B., Zwinderman, A. H., Tuk, B., Kamphuisen, H. A., and Oberye, J. J. (2000). Analysis of a sleep-dependent neuronal feedback loop: the slow-wave microcontinuity of the EEG. IEEE Trans. Biomed. Eng. 47, 1185–1194. doi: 10.1109/10.867928
Kerous, B., Skola, F., and Liarokapis, F. (2018). EEG-based bci and video games: a progress report. Virtual Real. 22, 119–135. doi: 10.1007/s10055-017-0328-x
Ko, W., Jeon, E., Jeong, S., Phyo, J., and Suk, H.-I. (2021). A survey on deep learning-based short/zero-calibration approaches for EEG-based brain-computer interfaces. Front. Hum. Neurosci. 15:643386. doi: 10.3389/fnhum.2021.643386
Kochura, Y., Gordienko, Y., Taran, V., Gordienko, N., Rokovyi, A., Alienin, O., et al. (2019). “Batch size influence on performance of graphic and tensor processing units during training and inference phases,” in International Conference on Computer Science, Engineering and Education Applications (Cham: Springer), 658–668. doi: 10.1007/978-3-030-16621-2_61
Kochura, Y., Stirenko, S., Alienin, O., Novotarskiy, M., and Gordienko, Y. (2017a). “Comparative analysis of open source frameworks for machine learning with use case in single-threaded and multi-threaded modes,” in 12th Int. Scientific and Technical Conf. on Computer Sciences and Information Technologies (CSIT) (IEEE), 373–376. doi: 10.1109/STC-CSIT.2017.8098808
Kochura, Y., Stirenko, S., Alienin, O., Novotarskiy, M., and Gordienko, Y. (2017b). “Performance analysis of open source machine learning frameworks for various parameters in single-threaded and multi-threaded modes,” in Conference on Computer Science and Information Technologies (Cham: Springer), 243–256. doi: 10.1007/978-3-319-70581-1_17
Koelstra, S., Muhl, C., Soleymani, M., Lee, J.-S., Yazdani, A., Ebrahimi, T., et al. (2011). Deap: a database for emotion analysis; using physiological signals. IEEE Trans. Affect. Comput. 3, 18–31. doi: 10.1109/T-AFFC.2011.15
Kostiukevych, K., Stirenko, S., Gordienko, N., Rokovyi, O., Alienin, O., and Gordienko, Y. (2021). “Convolutional and recurrent neural networks for physical action forecasting by brain-computer interface,” in 2021 11th IEEE International Conference on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications (IDAACS) (IEEE), 973–978. doi: 10.1109/IDAACS53288.2021.9660880
Kotowski, K., Stapor, K., and Ochab, J. (2020). “Deep learning methods in electroencephalography,” in Machine Learning Paradigms (Cham: Springer), 191–212. doi: 10.1007/978-3-030-49724-8_8
Lashgari, E., Liang, D., and Maoz, U. (2020). Data augmentation for deep-learning-based electroencephalography. J. Neurosci. Methods 346:108885. doi: 10.1016/j.jneumeth.2020.108885
Lashgari, E., Ott, J., Connelly, A., Baldi, P., and Maoz, U. (2021). An end-to-end cnn with attentional mechanism applied to raw EEG in a bci classification task. J. Neural Eng. 18:0460e03. doi: 10.1088/1741-2552/ac1ade
Lawhern, V. J., Solon, A. J., Waytowich, N. R., Gordon, S. M., Hung, C. P., and Lance, B. J. (2018). EEGnet: a compact convolutional neural network for EEG-based brain-computer interfaces. J. Neural Eng. 15:056013. doi: 10.1088/1741-2552/aace8c
LeCun, Y., Bengio, Y., and Hinton, G. (2015). Deep learning. Nature 521, 436–444. doi: 10.1038/nature14539
Lee, B.-H., Cho, J.-H., Kwon, B.-H., Lee, M., and Lee, S.-W. (2024). Iteratively calibratable network for reliable EEG-based robotic arm control. IEEE Trans. Neural Syst. Rehabil. Eng. 32:2793–2804. doi: 10.1109/TNSRE.2024.3434983
Lee, M.-H., Kwon, O.-Y., Kim, Y.-J., Kim, H.-K., Lee, Y.-E., Williamson, J., et al. (2019). EEG dataset and openbmi toolbox for three bci paradigms: an investigation into bci illiteracy. GigaScience 8:giz002. doi: 10.1093/gigascience/giz002
Leeb, R., Brunner, C., Müller-Putz, G., Schlögl, A., and Pfurtscheller, G. (2008). BCI competition 2008-graz data set b. Graz Univ. Technol. 16, 1–6. Available at: https://www.researchgate.net/publication/238115253_BCI_Competition_2008_Graz_data_set_B
Li, G., Lee, C. H., Jung, J. J., Youn, Y. C., and Camacho, D. (2020). Deep learning for EEG data analytics: a survey. Concurr. Comput. 32:e5199. doi: 10.1002/cpe.5199
Li, Y., Zhang, X.-R., Zhang, B., Lei, M.-Y., Cui, W.-G., and Guo, Y.-Z. (2019). A channel-projection mixed-scale convolutional neural network for motor imagery EEG decoding. IEEE Trans. Neural Syst. Rehabilit. Eng. 27, 1170–1180. doi: 10.1109/TNSRE.2019.2915621
Lin, B., Deng, S., Gao, H., and Yin, J. (2020). A multi-scale activity transition network for data translation in EEG signals decoding. IEEE/ACM Trans. Comput. Biol. Bioinform. 18, 1699–1709. doi: 10.1109/TCBB.2020.3024228
Lopez, E., Chiarantano, E., Grassucci, E., and Comminiello, D. (2023). “Hypercomplex multimodal emotion recognition from EEG and peripheral physiological signals,” in 2023 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW) (IEEE), 1–5. doi: 10.1109/ICASSPW59220.2023.10193329
Luciw, M. D., Jarocka, E., and Edin, B. B. (2014). Multi-channel EEG recordings during 3,936 grasp and lift trials with varying weight and friction. Sci. Data 1, 1–11. doi: 10.1038/sdata.2014.47
Micucci, D., Mobilio, M., and Napoletano, P. (2017). Unimib shar: a dataset for human activity recognition using acceleration data from smartphones. Appl. Sci. 7:1101. doi: 10.3390/app7101101
Nisar, M. A., Shirahama, K., Li, F., Huang, X., and Grzegorzek, M. (2020). Rank pooling approach for wearable sensor-based adls recognition. Sensors 20:3463. doi: 10.3390/s20123463
Okafor, E., Smit, R., Schomaker, L., and Wiering, M. (2017). “Operational data augmentation in classifying single aerial images of animals,” in 2017 IEEE International Conference on INnovations in Intelligent SysTems and Applications (INISTA) (IEEE), 354–360. doi: 10.1109/INISTA.2017.8001185
Ouyang, J., Wu, M., Li, X., Deng, H., and Wu, D. (2024). Briedge: EEG-adaptive edge ai for multi-brain to multi-robot interaction. arXiv preprint arXiv:2403.15432.
Pancholi, S., Giri, A., Jain, A., Kumar, L., and Roy, S. (2021). Source aware deep learning framework for hand kinematic reconstruction using EEG signal. arXiv preprint arXiv:2103.13862.
Parvan, M., Ghiasi, A. R., Rezaii, T. Y., and Farzamnia, A. (2019). “Transfer learning based motor imagery classification using convolutional neural networks,” in 2019 27th Iranian Conference on Electrical Engineering (ICEE) (IEEE), 1825–1828. doi: 10.1109/IranianCEE.2019.8786636
Peng, C.-K., Buldyrev, S., Goldberger, A., Havlin, S., Simons, M., and Stanley, H. (1993). Finite-size effects on long-range correlations: implications for analyzing dna sequences. Phys. Rev. E 47:3730. doi: 10.1103/PhysRevE.47.3730
Peng, C.-K., Havlin, S., Stanley, H. E., and Goldberger, A. L. (1995). Quantification of scaling exponents and crossover phenomena in nonstationary heartbeat time series. Chaos 5, 82–87. doi: 10.1063/1.166141
Pizzolato, S., Tagliapietra, L., Cognolato, M., Reggiani, M., Müller, H., and Atzori, M. (2017). Comparison of six electromyography acquisition setups on hand movement classification tasks. PLoS ONE 12:e0186132. doi: 10.1371/journal.pone.0186132
Rommel, C., Paillard, J., Moreau, T., and Gramfort, A. (2022). Data augmentation for learning predictive models on EEG: a systematic comparison. arXiv preprint arXiv:2206.14483.
Roy, Y., Banville, H., Albuquerque, I., Gramfort, A., Falk, T. H., and Faubert, J. (2019). Deep learning-based electroencephalography analysis: a systematic review. J. Neural Eng. 16:051001. doi: 10.1088/1741-2552/ab260c
Sakai, A., Minoda, Y., and Morikawa, K. (2017). “Data augmentation methods for machine-learning-based classification of bio-signals,” in 2017 10th Biomedical Engineering International Conference (BMEiCON) (IEEE), 1–4. doi: 10.1109/BMEiCON.2017.8229109
Simonyan, K., and Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556.
Soleymani, M., Lichtenauer, J., Pun, T., and Pantic, M. (2011). A multimodal database for affect recognition and implicit tagging. IEEE Trans. Affect. Comput. 3, 42–55. doi: 10.1109/T-AFFC.2011.25
Sun, C., and Mou, C. (2023). Survey on the research direction of EEG-based signal processing. Front. Neurosci. 17:1203059. doi: 10.3389/fnins.2023.1203059
Sun, J., Shen, A., Sun, Y., Chen, X., Li, Y., Gao, X., et al. (2024). Adaptive spatiotemporal encoding network for cognitive assessment using resting state EEG. NPJ Digital Med. 7:375. doi: 10.1038/s41746-024-01384-2
Talavera, E., Iglesias, G., González-Prieto, Á., Mozo, A., and Gómez-Canaval, S. (2022). Data augmentation techniques in time series domain: A survey and taxonomy. arXiv preprint arXiv:2206.13508.
Tangermann, M., Müller, K.-R., Aertsen, A., Birbaumer, N., Braun, C., Brunner, C., et al. (2012). Review of the bci competition IV. Front. Neurosci. 6:55. doi: 10.3389/fnins.2012.00055
Taran, V., Alienin, O., Stirenko, S., Gordienko, Y., and Rojbi, A. (2017). “Performance evaluation of distributed computing environments with hadoop and spark frameworks,” in 2017 IEEE Int. Young Scientists Forum on Applied Physics and Engineering (YSF) (IEEE), 80–83. doi: 10.1109/YSF.2017.8126655
Tsinganos, P., Cornelis, B., Cornelis, J., Jansen, B., and Skodras, A. (2018). “Deep learning in emg-based gesture recognition,” in PhyCS, 107–114. doi: 10.5220/0006960201070114
Tunnell, M., Chung, H., and Chang, Y. (2022). “A novel convolutional neural network for emotion recognition using neurophysiological signals,” in 2022 International Conference on Robotics and Automation (ICRA) (IEEE), 792–797. doi: 10.1109/ICRA46639.2022.9811868
Vahid, A., Mückschel, M., Stober, S., Stock, A.-K., and Beste, C. (2020). Applying deep learning to single-trial EEG data provides evidence for complementary theories on action control. Commun. Biol. 3, 1–11. doi: 10.1038/s42003-020-0846-z
Vasiljevic, G. A. M., and de Miranda, L. C. (2020). Brain-computer interface games based on consumer-grade EEG devices: a systematic literature review. Int. J. Hum. Comput. Interact. 36, 105–142. doi: 10.1080/10447318.2019.1612213
Wan, X., Zhang, K., Ramkumar, S., Deny, J., Emayavaramban, G., Ramkumar, M. S., et al. (2019). A review on electroencephalogram based brain computer interface for elderly disabled. IEEE Access 7, 36380–36387. doi: 10.1109/ACCESS.2019.2903235
Wang, F., Zhong, S.-,h., Peng, J., Jiang, J., and Liu, Y. (2018a). “Data augmentation for EEG-based emotion recognition with deep convolutional neural networks,” in MultiMedia Modeling: 24th International Conference, MMM 2018, Bangkok, Thailand, February 5–7, 2018, Proceedings, Part II 24 (Springer), 82–93. doi: 10.1007/978-3-319-73600-6_8
Wang, I.-N., Lee, C.-H., Kim, H., and Kim, D.-J. (2024). Negative-sample-free contrastive self-supervised learning for electroencephalogram-based motor imagery classification. IEEE Access. 12, 132714–132728. doi: 10.1109/ACCESS.2024.3459866
Wang, N., Farhadi, A., Rao, R., and Brunton, B. (2018b). “Ajile movement prediction: Multimodal deep learning for natural human neural recordings and video,” in Proceedings of the AAAI Conference on Artificial Intelligence, 2524–2531. doi: 10.1609/aaai.v32i1.11889
Wei, M., and Lin, F. (2020). A novel multi-dimensional features fusion algorithm for the EEG signal recognition of brain's sensorimotor region activated tasks. Int. J. Intell. Comput. Cybern. 13, 239–260. doi: 10.1108/IJICC-02-2020-0019
Wu, D., Li, S., Yang, J., and Sawan, M. (2022). neuro2vec: masked fourier spectrum prediction for neurophysiological representation learning. arXiv preprint arXiv:2204.12440.
Xia, T., Han, J., and Mascolo, C. (2022). “Benchmarking uncertainty quantification on biosignal classification tasks under dataset shift,” in Multimodal AI in Healthcare: A Paradigm Shift in Health Intelligence (Springer), 347–359. doi: 10.1007/978-3-031-14771-5_25
Xu, J., and Zhong, B. (2018). Review on portable EEG technology in educational research. Comput. Human Behav. 81, 340–349. doi: 10.1016/j.chb.2017.12.037
Zabcikova, M., Koudelkova, Z., Jasek, R., and Lorenzo Navarro, J. J. (2022). Recent advances and current trends in brain-computer interface research and their applications. Int. J. Dev. Neurosci. 82, 107–123. doi: 10.1002/jdn.10166
Zhang, K., Xu, G., Han, Z., Ma, K., Zheng, X., Chen, L., et al. (2020). Data augmentation for motor imagery signal classification based on a hybrid neural network. Sensors 20:4485. doi: 10.3390/s20164485
Zhang, Q., and Liu, Y. (2018). Improving brain computer interface performance by data augmentation with conditional deep convolutional generative adversarial networks. arXiv preprint arXiv:1806.07108.
Zhang, Y., Liao, Y., Zhang, Y., and Huang, L. (2021). Emergency braking intention detect system based on k-order propagation number algorithm: a network perspective. Brain Sci. 11:1424. doi: 10.3390/brainsci11111424
Zhang, Z., Huang, Y., Chen, S., Qu, J., Pan, X., Yu, T., et al. (2017). An intention-driven semi-autonomous intelligent robotic system for drinking. Front. Neurorobot. 11:48. doi: 10.3389/fnbot.2017.00048
Keywords: deep neural network, brain-computer interface, grasp-and-lift, physical action, data augmentation, noise, noise data augmentation, detrended fluctuation analysis
Citation: Gordienko Y, Gordienko N, Taran V, Rojbi A, Telenyk S and Stirenko S (2025) Effect of natural and synthetic noise data augmentation on physical action classification by brain–computer interface and deep learning. Front. Neuroinform. 19:1521805. doi: 10.3389/fninf.2025.1521805
Received: 02 November 2024; Accepted: 13 February 2025;
Published: 27 February 2025.
Edited by:
A. Amalin Prince, Birla Institute of Technology and Science, IndiaReviewed by:
Anurag Nishad, Birla Institute of Technology and Science, IndiaSaif Nalband, Thapar Institute of Engineering & Technology, India
Copyright © 2025 Gordienko, Gordienko, Taran, Rojbi, Telenyk and Stirenko. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Yuri Gordienko, Z29yZEBjb21zeXMua3BpLnVh