ORIGINAL RESEARCH article

Front. Hum. Neurosci., 19 January 2023

Sec. Brain-Computer Interfaces

Volume 16 - 2022 | https://doi.org/10.3389/fnhum.2022.1029784

Explainable artificial intelligence model to predict brain states from fNIRS signals

  • 1. Department of Computer Science, University of Arizona, Tucson, AZ, United States

  • 2. Division of Artificial Internal Organs, Department of Medical Devices Engineering, Biomedical Technology Wing, Sree Chitra Tirunal Institute for Medical Sciences and Technology, Trivandrum, India

  • 3. Department of Imaging Sciences and Interventional Radiology, Sree Chitra Tirunal Institute for Medical Sciences and Technology, Trivandrum, India

  • 4. Department of Diagnostic Imaging, St. Jude Children’s Research Hospital, Memphis, TN, United States

Article metrics

View details

21

Citations

6,7k

Views

2k

Downloads

Abstract

Objective: Most Deep Learning (DL) methods for the classification of functional Near-Infrared Spectroscopy (fNIRS) signals do so without explaining which features contribute to the classification of a task or imagery. An explainable artificial intelligence (xAI) system that can decompose the Deep Learning mode’s output onto the input variables for fNIRS signals is described here.

Approach: We propose an xAI-fNIRS system that consists of a classification module and an explanation module. The classification module consists of two separately trained sliding window-based classifiers, namely, (i) 1-D Convolutional Neural Network (CNN); and (ii) Long Short-Term Memory (LSTM). The explanation module uses SHAP (SHapley Additive exPlanations) to explain the CNN model’s output in terms of the model’s input.

Main results: We observed that the classification module was able to classify two types of datasets: (a) Motor task (MT), acquired from three subjects; and (b) Motor imagery (MI), acquired from 29 subjects, with an accuracy of over 96% for both CNN and LSTM models. The explanation module was able to identify the channels contributing the most to the classification of MI or MT and therefore identify the channel locations and whether they correspond to oxy- or deoxy-hemoglobin levels in those locations.

Significance: The xAI-fNIRS system can distinguish between the brain states related to overt and covert motor imagery from fNIRS signals with high classification accuracy and is able to explain the signal features that discriminate between the brain states of interest.

Introduction

Brain-Computer Interfaces (BCIs) provide communication pathways between the brain and external devices to be controlled (Wolpaw et al., 2002). BCIs have significant use in motor rehabilitation and communication and can help persons whose motor function or communication is compromised or who cannot communicate with the outside world (Sitaram et al., 2017). Furthermore, providing feedback of brain activity related to specific mental states may help the user control the BCI through operant learning or other forms of learning. The choice of the brain signals acquired and the features selected for feedback form essential steps in BCI development (Benitez-Andonegui et al., 2020; Rieke et al., 2020). BCIs may use electrical and magnetic signals acquired from Electroencephalography (EEG; Stojic and Chau, 2020; Gaur et al., 2021a, b), Magnetoencephalography (MEG; Roy et al., 2020; Ovchinnikova et al., 2021; Rathee et al., 2021), Electrocorticography (ECoG; Luo, 2020; Zhang et al., 2020), or implanted electrodes, or hemodynamic signals acquired from functional Magnetic resonance Imaging (fMRI; Simon et al., 2020; Sorger and Goebel, 2020), or functional near-infrared spectroscopy (fNIRS; Almulla et al., 2020; Li et al., 2020; Nazeer et al., 2020; Shibu et al., 2020; Ghaffar et al., 2021).

fNIRS measures hemodynamic activity by quantifying changes in the concentration of hemoglobin in the brain based on optical intensity measurements. With multiple probes over the subject’s entire scalp, fNIRS can detect hemodynamic activations in the superficial layers of the brain, thus allowing cognitive state determination using techniques such as pattern classification. This neuroimaging modality has recently been utilized to analyze brain functions in both healthy and diseased conditions (Arenth et al., 2007; Irani et al., 2007; Arun et al., 2018). fNIRS is a suitable signal acquisition system for BCI applications owing to its portability, ease of setting up, and robustness to artifacts caused by head motion (Naseer and Hong, 2015).

Coyle et al. (2007) were one of the first to validate the control of a binary switch using mental imagery based fNIRS BCI with an accuracy of above 80% for three subjects (Hong et al., 2015). Sitaram and coworkers could classify the fNIRS signals for motor imagery and motor execution with an accuracy of 80% for five healthy volunteers (Sitaram et al., 2007). Over the last decade, there has been increasing interest in fNIRS as the signal of choice for developing BCI (Naseer and Hong, 2013; Hong et al., 2015; Buccino et al., 2016; Naseer et al., 2016).

One of the broad aims of BCI research is to develop methods that improve the accuracy of brain state classification and ease of implementation. For the past decade, fNIRS-based BCIs primarily focused on extracting informative features from signals and implementing machine learning algorithms like K-Nearest Neighbor (KNN) and Support Vector Machine (SVM). The features included statistical properties like mean (Holper and Wolf, 2011; Faress and Chau, 2013; Naseer and Hong, 2013), signal peak (Bauernfeind et al., 2011; Holper and Wolf, 2011), signal slope (Naseer and Hong, 2013; Hong and Santosa, 2016), signal skewness (Tai and Chau, 2009; Holper and Wolf, 2011), signal kurtosis (Tai and Chau, 2009; Holper and Wolf, 2011), of the fNIRS time domain signals. Filter coefficients and discrete wavelet transforms (DWTs) have also been used (Khoa and Nakagawa, 2008; Abibullaev and An, 2012).

However, the extraction of these statistical features alone may limit classification performance as it depends on the types of features extracted which may not be an exhaustive set of discriminating features. Such manual feature selection is obviated by deep learning techniques which directly process the raw data. Deep learning has gained immense popularity due to its ability to automate feature selection and extraction processes (Chiarelli et al., 2018; Trakoolwilaiwan et al., 2018; Tanveer et al., 2019; Janani et al., 2020).

Deep learning models, along with their ability to extract features and learn from those features, are being used for various applications like classification, detection, and segmentation of images. Convolutional Neural Networks (CNN) are trained commonly for image classification (O’Shea and Nash, 2015; Nguyen et al., 2019; Olmos et al., 2019) and consist of convolutional layers for feature extraction from images. With multiple, fully connected layers, these models can learn to distinguish features of labeled images. Long short-term memory (Hochreiter and Schmidhuber, 1997) is a variant of the recurrent neural network and is designed specifically for time-series data. LSTM has the ability to remember important information for longer periods of time, and hence is most commonly used for sequence prediction, like speech recognition (Ying et al., 2020), human activity recognition (Wang and Liu, 2020), etc.

Trakoolwilaiwan et al. performed a study comparing the classification performance of SVM, Artificial Neural Networks (ANN), and Convolution Neural Networks (CNN). In the above work, the input data to the CNN consisted of changes in HbO- (oxyhemoglobin) and HbR-(deoxyhemoglobin) concentration of all channels available with a dimension of M by N where M is the number of data points and N is the number of channels available. Various combinations of 1D CNNs (single and multiple convolutional layers, and varying the number of filter kernels) achieving an average classification of over 89% for discriminating rest, left, and right states (Trakoolwilaiwan et al., 2018).

Nagabushanam et al. (2020) compared Support Vector Machine (SVM) and Long-Short Term Memory (LSTM) classification techniques from EEG signals for various tasks and showed that LSTM performs better than SVM by roughly 15%.

A recent study by Z. Shi et al. (Lu et al., 2020) comparing machine learning models like SVM, KNN, and LDA with deep learning models like long short-term memory-fully convolutional network (LSTM-FCN), found that deep learning models could achieve a decode mental arithmetic task with a classification accuracy of 97%.

In a recent study by Ghonchi et al. (2020), spatiotemporal maps were extracted from fNIRS-EEG signals during a motor imagery task to classify the task conditions by a deep neural network. The authors were able to classify the brain states of motor imagery with an accuracy of 99%.

With the growing ability of complex yet reliable Deep Learning models to classify brain states there seems to be a tradeoff between with accuracy and interpretability of a model’s output. Deep Learning network was in the past a black box, because an explanation in terms of which features affected the trained model to arrive at a specific decision was missing.

The eXplainable Artificial Intelligence (XAI) developed in recent years aims to explain the supervised machine learning models (Gunning and Aha, 2019). A few XAI models in the recent literature are summarized below:

  • 1.

    Local Interpretable Model-Agnostic Explanations (LIME; Ribeiro et al., 2016) attempts to comprehend the models by varying the input data samples and analyze the changes in predictions. LIME modifies a single input value by changing the feature value and then observes the result. LIME generates a list of explanations that indicate the contribution of each characteristic to the data sample prediction. LIME has a few drawbacks, such as the fact that its explanations are not resilient, meaning that a minor change in the input causes the explanations to alter substantially the importance of features (Alvarez-Melis and Jaakkola, 2018).

  • 2.

    Deep Learning Important FeaTures (DeepLIFT) DeepLIFT aims to explain the difference between a neural reference and an input (Shrikumar et al., 2017). DeepLIFT is a model explanation that gives input variables a significant score. This score is determined by building a partial derivative-like function that is used in conjunction with the chain rule to track the change in the output layer relative to the input layer.

  • 3.

    SHAP (SHapley Additive exPlanations) is a method that uses shapley values to explain the model’s output (Lundberg and Lee, 2017). Shapley values are a concept acquired from game theory where different players in a coalition are analyzed for each player’s contribution to the coalition’s success. DeepSHAP is a fast approximation approach for computing SHAP values in DeepLIFT’s built-in deep learning models. It uses feature significance determined using linear composition rules to explain the output, and the chain rule to back propagate the activation difference from the output layer to the original input.

Given its success in recent literature, we propose a SHAP-based explainable AI-fNIRS (xAI-fNIRS) system in our study. We will use this approach: (i) to classify fNIRS signals obtained during overt motor execution tasks as well as covert motor imagery tasks to accurately predict the task; and (ii) to find the contribution of each channel to the model’s performance during classification. The proposed xAI-fNIRS system shown in Figure 1, comprises of a classification Deep Learning model for classifying brain states, during overt motor execution and covert motor imagery conditions which is generalized over two different datasets, followed by an explanation module. The latter consists of DeepSHAP network to introduce explainability to the model output by constructing a decomposition of CNN output on the input variables (i.e., channels).

Figure 1

Methods

Data acquisition

Dataset A

Three healthy subjects (Two females and one male) with a mean age of 25 years participated in the study. Ethical clearance was obtained for data acquisition from the Institutional Ethics committee of Sree Chitra Tirunal Institute for Medical Sciences and Technology ethics committee. For data acquisition, a continuous wave, multichannel fNIRS device (NIRsport from NIRx Medical Technologies LLC, Berlin, Germany, RRID: SCR_002491 and NIRStar (v.15.2) software (from NIRx Medical Technologies LLC, Berlin, Germany, RRID: SCR_014540) with two wavelengths (750 nm and 830 nm) and a sampling frequency of 7.8125 Hz, was employed. A total of eight sources and eight detectors were used in combination to form 20 channels. They were placed over bilateral motor cortices for signal acquisition, as shown in Figure 2. The Optode locations are based on the international 10–20 system (Homan et al., 1987). The distance between the source and the detector was 3 cm.

Figure 2

Dataset B

We used an open-access dataset from Technische Universität Berlin containing fNIRS motor imagery data. The dataset contained 28 right-handed and one healthy left-handed subjects with average age (years) 28.5 ± 3.7.

There were no neurological, neuropsychiatric, or other brain-related illnesses reported by any of the subjects. The experimental protocol was explained to all volunteers, and informed consent was obtained from all participants (Shin et al., 2017).

At the sampling rate of 10 Hz NIRS data was acquired using NIRScout (NIRx GmbH, Berlin, Germany). One physiological NIRS channel is created by each nearby source-detector pair. At three brain areas, 14 sources and 16 detectors were used to create 36 physiological channels. The frontal (nine channels around Fp1, Fp2, and Fpz), motor (12 channels around C3 and C4, respectively), and visual (three channels around Oz) portions of the brain were studied. (Figure 3).

Figure 3

Experimental procedure

Dataset A

The subjects were asked to be seated on a chair facing a computer monitor on which the experimental task was displayed. To generate a robust signal of the neural activity, subjects were requested to perform a motor execution. Subjects were instructed to relax before the paradigm began, to stabilize blood flow to the cortex. They were then instructed to relax during the rest period and open and close their left-hand or right-hand palm depending on the screen’s cue. Each subject had four sessions, and each session had 10 rest periods, each following an experimental event. The experiment involved 10 active motor execution tasks (five left and five right), each of 20-s duration separated by 20 s rest periods (Figure 4).

Figure 4

Dataset B

Subjects were asked to engage in kinesthetic motor imagery (imagining opening and closing their hand while grasping a ball). A visual instruction appears on the screen in the form of a black arrow pointing left or right for 2 s. Throughout the activity, MI was performed for 10 s at a time, followed by a 15–17 s rest interval. In a single session, this was done 20 times (10 trials for each left- and right-hand MI in a single session: 30 trials for each one in the whole three sessions). MI tasks were done in a single session based on 10 following blocks, each consisting of one of two conditions: left-hand MI first, then right-hand MI, or vice versa (Shin et al., 2017; Figure 5).

Figure 5

Signal pre-processing

AnalyzIR Toolbox (Santosa et al., 2018) was used to pre-process the fNIRS data. The raw signals were passed through a bandpass filter with a passband of 0.01–0.2 Hz. The filter helps in removing cardiac, respiratory, and other motion artifacts from the signal. Filtered signals were converted to oxygenated (ΔcHbO(t)) and deoxygenated (ΔcHbR(t)) hemoglobin concentration changes using modified Beer-Lamberts Law (Baker et al., 2014):

where ΔA(t,λj), j represents unitless absorbance (optical density), which is a measurement of the variation in light emitted at a specific wavelength. λj, αcHbO1) and αcHbR1) are the extinction coefficients of HbO and HbR (μ−1 mm−1) respectively, d is the unitless differential path length factor (DPF), and l is the distance between the emitter and detector. With the distance between the source and detector by default was maintained at 30 mm, NIRS signal is able to penetrate 15 mm into the gray matter of the brain (Gratton et al., 2006; Zhou et al., 2020; Yücel et al., 2021). Because fNIRS signals are sensitive to motion artifacts like respiration and heartbeat, channels having a coefficient of variation (CV, the standard deviation/mean) of each wavelength exceeding 7.5 were discarded (Minati et al., 2011; Balardin et al., 2017; Pfeifer et al., 2018).

Background

Convolutional neural network

The CNN architecture shown in Figure 6 can be divided into four parts (O’Shea and Nash, 2015):

Figure 6

  • 1.

    The input layer consists of an input matrix, which in our case, on the x-axis represents the fNIRS channels and the y-axis represents the fNIRS time-series data from each channel after windowing.

    The convolutional layer has 32 1D kernels of size 3 (span in time, and spans across all channels), which calculates the neurons’ output for each channel.

    The learnable kernels (filters used to extract features from the image matrix) use a small spatial dimensionality. When the data arrives at the convolutional layer, it convolves each filter across the input’s temporal dimension to produce a 2D feature map. The kernels have a corresponding activation map. Convolutional layers are optimized through three hyper-parameters, namely depth, stride, and zero-padding (Albawi et al., 2017). Depth of an image (assume a color image of dimension 50 × 50 and a depth of 3 which is RGB) or a grayscale video (dimension is height and breadth, and depth are the frames). Each convolutional layer has its own filter of a certain dimension which slides over the input matrix, this is controlled by stride. Stride is the number of pixels shifts over the input matrix. Zero padding is a commonly used modification where symmetrically zeros are added to the input matrix. This is done to preserve the dimension of the input volume in the output volume.

  • 2.

    The pooling layer down samples along the temporal dimension, reducing the number of parameters within that feature map.

  • 3.

    The fully connected layers contain neurons directly connected to adjacent layers without connecting to any of the neurons within the same layer. These three layers produce a class score from the activation maps, which are then used for classification.

Long short-term memory

  • Long Short-Term memory, shown in Figure 7, is a type of Recurrent Neural Network (RNN) architecture that allows learning from long-term dependencies (Hochreiter and Schmidhuber, 1997; Sutskever et al., 2014; Lipton et al., 2015). The central concept of LSTM is a memory cell that maintains its state over time and is made up of explicit memory (cell state vector) and gating units that control information flow into and out of the memory.

  • The LSTM’s memory is represented by the cell state vector, which alters the forgetting of old memory (forget gate) and the addition of new memory (input gate). The output from the previous time step, the current input, and, optionally, the cell status vector are used to govern the flow of information to and from memory.

Figure 7

Classification metrics

This section explains the strategy used to evaluate the model’s performance using Accuracy, Precision, Recall, and F1 score (Goutte and Gaussier, 2005; Lin et al., 2018; Powers, 2020). First, a brief definition of key terms:

False-positive (FP): the rejection of true null hypothesis; False negative (FN): acceptance of the false null hypothesis; True positive (TP): outcome of the model that correctly predicts observation belonging to the positive class. True negative (TN): outcome of the model that correctly predicts observation belonging to the negative class. Accuracy, precision, recall, and F1 score are computed using these as follows:

Proposed sliding window approach

A sliding window looks at the latest samples from the signal. fNIRS data is recorded as a time-series with M time samples and N channels where X1 is the latest block of size WS, and this block slides one row in time (from row1 to rowM-WS) till XM-WS. This gives an output of dimension [(M-WS) × WS × N] as shown in Figure 8.

Figure 8

For our study, we selected a window size of 50, which represents roughly 6 s of data for dataset A and 5 s for dataset B (open access). From the sliding window function’s output, each block’s dimension was computed as 1 × WS × N which is shown in Figure 9 where time is the WS and N is number of channels. Finally, this is labeled as “left,” “rest,” or “right” based on the experimental paradigm.

Figure 9

We followed two approaches to label each block:

  • Labeling approach 1: Consider a block of specific window size, WS. If, for example, the label “rest” appears in more than 50% of the latest block, then that block is labeled as “rest.” The same method is applied to label blocks as “right” or “left.” For Dataset A this approach has a latency of 4 s, and for Dataset B it has a latency of 5 s.

  • Labeling approach 2: Consider a block of window size (WS). If all the latest 1 s of labels for this block have a specific label, say left, then that block is labeled as left Likewise, the same is done in case the latest 1 s is labeled as rest or right. For both Dataset A and Dataset B, this approach has a latency of 1 s.

Proposed structure

For better understanding of the CNN and LSTM models, the input and output sizes of each layer in our proposed CNN model are summarized in Table 1 and likewise for LSTM model are summarized in Table 2. Relu is an activation function that stands for rectified linear unit. If the input is positive, this piecewise linear function will output the input directly; otherwise, it will output a zero (Goodfellow et al., 2016; Arora et al., 2018; Agarap, 2019). Because it is faster to train the model and generally produces high accuracy, most neural networks utilize relu as their default activation function. Softmax is also an activation used in the output layer of the neural network that predicts a multinomial probability distribution for a multi-class classification problem (Goodfellow et al., 2016). Both Dataset A and B used “relu” as the activation function for all layers except the fully connected layer 3 which used a “softmax” as its activation function. Adam is used as an optimization algorithm instead of the classical stochastic gradient descent approach to iteratively update the weights based in the training data (Kingma and Ba, 2017). We used Adam as an optimizer as it has various benefits like easy implementation, computational efficiency, less memory requirements, etc.

Table 1

LayerInput sizeOutput sizeProperties
Convolutional Layer 150,2650,3232 filters with kernel size 3, strides 1
Max Pooling layer 150,3228,32Pool size 2
Flatten layer 128,32896
Fully connected layer 1896500500 hidden nodes
Dropout layer 1500500Dropout rate set to 50%
Fully connected layer 2500120120 hidden nodes
Dropout layer 2120120Dropout rate set to 50%
Fully connected layer 312033 hidden nodes

Input and output shapes of the Convolutional Neural Network of this study.

Table 2

LayerInput sizeOutput sizeProperties
LSTM layer50,2650,120120 LSTM neurons
Flatten layer 150,1206,000
Fully connected layer 16,000500500 hidden nodes
Dropout layer 1500500Dropout rate set to 50%
Fully connected layer 2500120120 hidden nodes
Dropout layer 2120120Dropout rate set to 50%
Fully connected layer 312033 hidden nodes

Input and output shapes of Long Short-Term Memory Neural Network of this study.

DeepSHAP explanation module

It is computationally expensive to explore every potential feature combination while working with our datasets. DeepSHAP is a useful tool for estimating Shapley values in deep learning models. We are able to maintain the interpretive capability of Shapley values while maintaining the computational power and accurate outcomes of deep learning models. We can model Shapley values by additive feature attribution techniques for SHAP estimates. Through the explanatory model g, we describe the “coalition vector” z for SHAP as follows:

Here, we add the feature attribution for j, , multiplied by the coalition vector for j, , from feature j = 1 to the largest coalition size, M. () is the “coalition vector”. In the coalition vector, each feature is either “present” or “missing” in the combination of characteristics, denoted by a 1 or a 0.

We used DeepSHAP for our explanation module as it is optimized for explaining deep neural networks while having benefits which are:

  • (a)

    Global interpretability—The SHAP values can describe how each predictor or input variable contributes, either negatively or positively, to the target variable. This can be used to show each predictor’s positive and negative relationship with the target.

  • (b)

    Local interpretability—The SHAP value is assigned to each observation, which increases the transparency by showing how each case is predicted in terms of the contribution of each of the predictors. Traditional interpreters don’t explain individual cases but rather across the entire population. This enables us to pinpoint and contrast the impact of the predictors.

Using SHAP we can also find feature dependence, which is a form of global interpretation plot, where we choose a feature, and plot a point with the feature value on the x-axis and the appropriate Shapley value on the y-axis for each data instance.

Mathematically, we can describe the plot as:

With feature dependence we can also find SHAP interaction, which is interaction effects when pairwise features are attributions are considered. It is mathematically given as:

When ij:

Where S represents a subset of all features N except the i-th feature and M is the number of input features. This formula calculates the difference between the SHAP values of the i-th feature with and without the j-th feature and can be used to interpret the SHAP interaction value of the i-th feature with respect to the j-th feature. This enables us to compute SHAP interaction values using the algorithm for computing SHAP values (Lundberg and Lee, 2017).

Results

Hemodynamic response

The hemodynamic response obtained from subject 2 of Dataset A having 26 channels across all sessions of each task for rest, left, and right motor tasks with a window size of 50 are shown in Figure 10.

Figure 10

Changes in HbO- and HbR-concentration were measured as input to our proposed model.

Classification accuracies

LSTM and CNN were used to determine classification accuracies for each run of every subject. 10-fold cross-validation was performed on data with training- and validation- data split as 75% for training and 25% for testing.

The result of validation accuracy for each run of every subject for a window size WS of 50 samples (roughly 6 s of data) is shown in Figures 11 and 12 below:

Figure 11

Figure 12

The mean classification accuracy of CNN for Dataset A was 97.68% ± 0.05%, and that for Dataset B was 97.92% ± 0.28%. In LSTM, an average classification accuracy of 98.69% ± 0.04% was recorded for Dataset A and 97.88% ± 0.21% for Dataset B. To test the model for overfitting we randomized the labels. We then performed a permutation test with 10-fold cross-validation on Dataset A with randomized labels, which gave an accuracy of 39.54% ± 1.60% and 36.06% ± 1.88% for CNN and LSTM, respectively. Similarly, 10-fold cross-validated permutation test on Dataset B with randomized labels gave average accuracies of 56.55% ± 0.33% for CNN and 56.62% ± 0.39% for LSTM. Therefore our model is robust.

Subject independent classification

For subject independent classification we combined a few subjects (a combination of three, five and 10) for training and leaving out one for testing with 10-fold cross validation. The results are shown in Table 3. Dataset A gave an average classification accuracy of 50%. The same couldn’t be done for Dataset B as the number of channels from one subject to another varied.

Table 3

ModelTraining with subjectsTesting with subjectsTesting Accuracy
Labeling Approach 1Labeling Approach 2
CNN1 and 2351.06% (+/– 0.09%)51.62% (+/– 0.11%)
LSTM49.53% (+/– 0.10%)47.98% (+/– 0.11%)
CNN1–3447.99% (+/– 0.10%)47.98% (+/– 0.08%)
LSTM48.20% (+/– 0.08%)44.32% (+/– 0.09%)
CNN1–4542.30% (+/– 0.11%)43.30% (+/– 0.07%)
LSTM46.05% (+/– 0.06%)42.13% (+/– 0.08%)
CNN1–5657.22% (+/– 0.08%)50.13% (+/– 0.12%)
LSTM53.72% (+/– 0.02%)48.52% (+/– 0.08%)
CNN1–6747.33% (+/– 0.10%)50.55% (+/– 0.08%)
LSTM50.20% (+/– 0.08%)46.19% (+/– 0.05%)
CNN1–7852.93% (+/– 0.06%)51.10% (+/– 0.11%)
LSTM53.29% (+/– 0.07%)52.27% (+/– 0.05%)
CNN1–8954.87% (+/– 0.08%)54.12% (+/– 0.09%)
LSTM56.78% (+/– 0.10%)53.02% (+/– 0.13%)
CNN1–91050.10% (+/– 0.09%)38.38% (+/– 0.07%)
LSTM48.22% (+/– 0.09%)42.23% (+/– 0.08%)

Subject independent classification.

Classification accuracy and validation loss

Figure 13 shows that our model is not overfitting for both CNN and LSTM as training- and validation- accuracies and losses plots converge for each epoch.

Figure 13

Visualizing the convolutional filters

Figure 14 shows the CNN model’s features for subject 3 of Dataset A, which is challenging to interpret. SHAP values were used to understand which features contributed the most. This is explained in the next section.

Figure 14

Explanations

Our studies involved a Deep Learning model, and this necessitated the use of an explainer module. In our study, we used DeepExplainer, a module in DeepSHAP. The DeepExplainer examined the CNN model along with 2,000 random samples of data from the training set. SHAP values were extracted from 1,000 random samples from the test dataset. We obtained a SHAP value of the dimension 3 × 1,000 × 50 × 72 where 3 is the number of classes and 1,000 is the number of random samples of dimension 50 frames of data from 72 channels.

Figure 15, for Dataset A, the 10 most important channels for each of the three brain states (Left active, right active, and Rest) are shown. Figure 16, for Dataset B, the 10 most important channels for each of the three brain states (LMI, RMI, and Rest) are shown. With respect to deThe Channels are ranked by decreasing mean relevance calculated from local, back-propagated relevance ratings over subject 1 for Dataset A and subject 1 for dataset B.

Figure 15

Figure 16

In Figures 15 and 16, the parameters are displayed as a blue horizontal bar in ascending order of global parameter importance, as determined by mean relevance. All of the individual data points are colored by parameter value and sorted by mean importance in the local explanation summary. The number of data-points at the related level of relevance correlates with the height of the data-points for each feature, as determined by back-propagated relevance scores for each feature. The feature value linked with the local explanation is used to color-code this. The influence of a feature on the model’s categorization is indicated by the position of the dot on the x-axis. Multiple dots that land at the same × location pile up to demonstrate density. For example, Figure 15B Right active (Motor task) seems to associate with the high FC3C3deoxy and low C1FC1oxy from the left cortex. Likewise, Figure 16B Right motor imagery seems to be associated with high AF4AFzdeoxy and C2C4deoxy from the right frontal cortex and right motor cortex. The summary distribution allows us to get an idea of what to expect from the fNIRS signal classification model.

Another interesting finding is shown in Figure 17A Left motor imagery and (B) Right motor imagery where high values of channel C6C4deoxy contribute to other classes (rest or motor left imagery) than itself (right motor imagery), but it contributes to Rest state in (C) where it clearly shows the high values for channel C6C4deoxy contribute more towards the classification of Rest state.

Figure 17

The dependence plot shows the marginal effect one or two features have on a Deep Learning model’s predicted outcome. Dependence plots display the relationship between a feature’s value (x-axis) and the prediction of each sample (each dot) in a dataset (y-axis). In comparison to conventional partial dependency charts, they offer deeper information. Figure 17 is a SHAP dependence plot where a feature that has a positive effect on classification is plotted against the value of that feature for all examples in the dataset.

Figure 17B right active, SHAP value of FC3C3deoxy is plotted against its actual value, clear interaction with FC3C3oxy, which increases the classification impact towards right active (Motor task). We can also observe that when the FC3C3deoxy value is low it has a negative impact and classifies anything other than right active, when the FC3C3deoxy value is high it has a positive impact towards classification of right active. Likewise, Figure 17D. Left motor imagery (LMI) channel CP4CP6deoxy shows interaction with channel FC2FC4deoxy, low values CP4CP6deoxy have negative impact towards classification of LMI and high values of CP4CP6deoxy contribute more towards LMI.

Discussion

Deep Learning models like CNN and LSTM enable automated feature extraction and classification of fNIRS signals. The previous works on fNIRS used statistical features which were extracted for machine learning algorithms (Naseer and Hong, 2013; Naseer et al., 2014), e.g., mean, slope, skewness, etc. of HbO and HbR, which yielded a classification accuracy of 86.19% for SVM whereas CNN enhanced the classification accuracy to 93.08% (Trakoolwilaiwan et al., 2018).

The window-based approach (Naseer and Hong, 2013) restricts attention only to variation in data for contiguous windowed periods, whereas the proposed sliding window approach looks at multiple overlapping windows starting with each input sample.

As data scarcity is one limiting factor of Deep Learning techniques (Ghonchi et al., 2020; Lu et al., 2020; Nagabushanam et al., 2020) the sliding window also augments the data by increasing the data’s sample size, as explained in Section “Proposed sliding window approach”.

The Supplementary Material shows the number of channels available channels in Dataset A, which shows that despite fewer channels, the classification module can classify with an accuracy over 98% accuracy for both CNN and LSTM shown in Figure 12. We used Dataset B, an open-access dataset (Abibullaev and An, 2012) that consisted of motor imagery for 29 subjects to bolster our study. The classification module also gave a classification accuracy of over 98% for both CNN and LSTM for Dataset B, shown in Figure 12.

We trained the model with various window sizes from 1 s to 8 s and 5–6 s (50 frames), which led to the most optimal results. This shows that the longer the sliding window’s duration, the greater the amount of historical data in the window, which contributes to better classification accuracy.

For Dataset A, we could not perform intersubjective group classification as the number of channels did not remain constant for all three subjects. As for Dataset B, we performed intersubjective group classification shown in Supplementary Material; we observe that the accuracy is low due to a large outlier or lack of normalization of optode placement across all subjects.

The spatiotemporal features identified in Figure 14 show the fNIRS signal variation, from which the model has learned features for classification. Each of the filters describes temporal variation for three-time samples and spatial variation along the 26 channels (subject 3 of dataset A); however, it is not interpretable.

To introduce interpretability and trackability in neural networks for BCI purposes so far, few EGG studies have shown that it is possible to represent using direct pixel-to-channel mapping (Joshi et al., 2018), projection of sensor locations by interpolating image of the sensor on the scalp (Bashivan et al., 2015), and an fNIRS study channels with a frequency greater than mean were considered to be most contributing channels another fNIRS study used z-score and t-value methods for channel selection (Nazeer et al., 2020).

SHAP (SHapley Additive exPlanations) is a novel method that employs a game-theoretical approach to describe the output of a Deep Learning model. A recent study used SHAP to explain which channels contributed the most in classifying EEG data for the Active touch task (Alsuradi et al., 2020).

In the explanation module of the “xAI-fNIRS” system we aimed to explain which features contribute to the classification of a specific task in the CNN model for both Dataset A and Dataset Busing DeepSHAP which is a module in SHAP.

From Figures 16 and 17, we observe the high-relevance features (Channels) for Dataset A and Dataset B. We plotted 10 features (channels) that show the highest level of relevance for classification in descending order. An interesting finding is that the explanation module showed that Dataset A had all oxy values of channels that contribute the most to motor task brain states’ classification, whereas Dataset B had all deoxy values of channels for the classification of motor imagery brain states.

It is important to note that the xAI-fNIRS system presented in this study should not be conceived as the one-and-only multi-outcome model. Instead, it should be viewed as a general model for building precise and explainable models towards the development of better BCI systems.

In summary, we have presented the xAI-fNIRS system to classify brain activity states, which can be used for both task and imagery tasks in fNIRS. The xAI-fNIRS shows high classification performance while enabling the possibility of explaining the classification in terms of pinpointing decisive input data to provide BCI researchers to understand the underlying reasons for classification. We hope that our study will lead to developing a better BCI system that could not only perform with high accuracy prediction of brain states but also help in understanding the underlying spatiotemporal activation patterns.

Conclusion

The xAI-fNIRS system’s classification module showed high classification accuracy of greater than 98% a sliding window approach with Deep Learning models like CNN and LSTM. The sliding window approach keeps a history of signals and, at the same time, increases the dimension of the dataset. The classification module performed well, giving a high classification accuracy of over 98% for two datasets (Motor task and Motor imagery). Deep learning has gotten so good at classifying brain data, but it fails to reveal which features contribute the most to classification. To solve this issue, the explanation module aims to bring insight into reasons for classification by every channel given a color-coded relevance score; a positive relevance score contributes to the true class and vice versa. Hence the xAI-fNIRS system is a step towards building better BCI systems.

Statements

Data availability statement

The raw data supporting the conclusion of this article can be made available on request. However, this is subject to the data sharing protocols of SCTIMST, Trivandrum, India.

Ethics statement

The studies involving human participants were reviewed and approved by Institutional Ethics Committee (IEC) of the Sree Chitra Tirunal Institute for Medical Sciences and Technology, Trivandrum. The patients/participants provided their written informed consent to participate in this study. Informed consent was obtained from all individual participants included in the study. The clinical trial has been registered with identification no.SCT/IEC/801/AUGUST-2015.

Author contributions

CJS developed the xAI part of the Deep Learning network. CJS and SS were involved in the design of the Deep Learning network. All authors were involved in the formulation of the methodology, contributed to the writing of the article and approved the submitted version.

Funding

This study was funded by the Department of Biotechnology (DBT/IN/New Indigo/02/CK/2013) and the Indian Council of Medical Research (ICMR SRF File NO.5/3/8/50/ITR-F/2018-ITR). RS was supported by Comisión Nacional de Investigación Científica y Tecnológica de Chile (Conicyt) through Fondo Nacional de Desarrollo Científico y Tecnológico, Fondecyt (project n° 1171320 and 1171313) and CONICYT-PIA Anillo ACT172121.

Acknowledgments

We thank the Director of Sree Chitra Tirunal Institute for Medical Sciences and Technology for providing the resources to conduct this research.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnhum.2022.1029784/full#supplementary-material.

References

  • 1

    AbibullaevB.AnJ. (2012). Classification of frontal cortex haemodynamic responses during cognitive tasks using wavelet transforms and machine learning algorithms. Med. Eng. Phys.34, 13941410. 10.1016/j.medengphy.2012.01.002

  • 2

    AgarapA. F. (2019). Deep learning using rectified linear units (ReLU). arXiv [Preprint]. 10.48550/arXiv.1803.08375

  • 3

    AlbawiS.MohammedT. A.Al-ZawiS. (2017). “Understanding of a convolutional neural network,” in 2017 International Conference on Engineering and Technology (ICET), (Antalya, Turkey), 16.

  • 4

    AlmullaL.Al-NaibI.AlthobaitiM. (2020). Hemodynamic responses during standing and sitting activities: a study toward fNIRS-BCI. Biomed. Phys. Eng. Express.6:055005. 10.1088/2057-1976/aba102

  • 5

    AlsuradiH.ParkW.EidM. (2020). “Explainable classification of EEG data for an active touch task using shapley values,” in HCI International 2020 - Late Breaking Papers: Multimodality and Intelligence6, eds StephanidisC.KurosuM.DegenH.Reinerman-JonesL. (Cham: Springer International Publishing), 406416.

  • 6

    Alvarez-MelisD.JaakkolaT. S. (2018). On the robustness of interpretability methods. arXiv [Preprint]. 10.48550/arXiv.1806.08049

  • 7

    ArenthP. M.RickerJ. H.SchultheisM. T. (2007). Applications of functional near-infrared spectroscopy (fNIRS) to neurorehabilitation of cognitive disabilities. Clin. Neuropsychol.21, 3857. 10.1080/13854040600878785

  • 8

    AroraR.BasuA.MianjyP.MukherjeeA. (2018). Understanding deep neural networks with rectified linear units. arXiv [Preprint]. 10.48550/arXiv.1611.01491

  • 9

    ArunK. M.SmithaK. A.RajeshP. G.KesavadasC. (2018). Functional near-infrared spectroscopy is in moderate accordance with functional MRI in determining lateralisation of frontal language areas. Neuroradiol. J.31, 133141. 10.1177/1971400917739083

  • 10

    BakerW. B.ParthasarathyA. B.BuschD. R.MesquitaR. C.GreenbergJ. H.YodhA. G. (2014). Modified Beer-Lambert law for blood flow. Biomed. Opt. Express5, 40534075. 10.1364/BOE.5.004053

  • 11

    BalardinJ. B.Zimeo MoraisG. A.FuruchoR. A.TrambaiolliL.VanzellaP.BiazoliC.et al. (2017). Imaging brain function with functional near-infrared spectroscopy in unconstrained environments. Front. Hum. Neurosci.11:258. 10.3389/fnhum.2017.00258

  • 12

    BashivanP.RishI.YeasinM.CodellaN. (2015). Learning representations from EEG with deep recurrent-convolutional neural networks. arXiv [Preprint]. 10.48550/arXiv.1511.06448

  • 13

    BauernfeindG.SchererR.PfurtschellerG.NeuperC. (2011). Single-trial classification of antagonistic oxyhemoglobin responses during mental arithmetic. Med. Biol. Eng. Comput.49, 979984. 10.1007/s11517-011-0792-5

  • 14

    Benitez-AndoneguiA.BurdenR.BenningR.MöckelR.LührsM.SorgerB. (2020). An augmented-reality fNIRS-based brain-computer interface: a proof-of-concept study. Front. Neurosci.14:346. 10.3389/fnins.2020.00346

  • 15

    BuccinoA. P.KelesH. O.OmurtagA. (2016). Hybrid EEG-fNIRS asynchronous brain-computer interface for multiple motor tasks. PLoS One11:e0146610. 10.1371/journal.pone.0146610

  • 16

    ChiarelliA. M.CroceP.MerlaA.ZappasodiF. (2018). Deep learning for hybrid EEG-fNIRS brain-computer interface: application to motor imagery classification. J. Neural Eng.15:036028. 10.1088/1741-2552/aaaf82

  • 17

    CoyleS. M.WardT. E.MarkhamC. M. (2007). Brain–computer interface using a simplified functional near-infrared spectroscopy system. J. Neural Eng.4, 219226. 10.1088/1741-2560/4/3/007

  • 18

    FaressA.ChauT. (2013). Towards a multimodal brain-computer interface: combining fNIRS and fTCD measurements to enable higher classification accuracy. Neuroimage77, 186194. 10.1016/j.neuroimage.2013.03.028

  • 19

    GaurP.GuptaH.ChowdhuryA.McCreadieK.PachoriR. B.WangH. (2021a). A sliding window common spatial pattern for enhancing motor imagery classification in EEG-BCI. IEEE Trans. Instrum. Meas.70, 19. 10.1109/TIM.2021.3051996

  • 20

    GaurP.McCreadieK.PachoriR. B.WangH.PrasadG. (2021b). An automatic subject specific channel selection method for enhancing motor imagery classification in EEG-BCI using correlation. Biomed. Signal Process. Control68:102574. 10.1016/j.bspc.2021.102574

  • 21

    GhaffarM. S. B. A.KhanU. S.IqbalJ.RashidN.HamzaA.QureshiW. S.et al. (2021). Improving classification performance of four class FNIRS-BCI using Mel Frequency Cepstral Coefficients (MFCC). Infrared Phys. Technol.112:103589. 10.1016/j.infrared.2020.103589

  • 22

    GhonchiH.FatehM.AbolghasemiV.FerdowsiS.RezvaniM. (2020). “Spatio-temporal deep learning for EEG-fNIRS brain computer interface,” in 2020 42nd Annual International Conference of the IEEE Engineering in Medicine Biology Society (EMBC), (Montreal, QC, Canada), 124127.

  • 23

    GoodfellowI.BengioY.CourvilleA. (2016). Deep Learning.Cambridge, MA: MIT Press.

  • 24

    GoutteC.GaussierE. (2005). “A Probabilistic interpretation of precision, recall and F-score, with implication for evaluation,” in Advances in Information Retrieval, eds LosadaD. E.Fernández-LunaJ. M. (Berlin, Heidelberg: Springer), 345359.

  • 25

    GrattonG.BrumbackC. R.GordonB. A.PearsonM. A.LowK. A.FabianiM. (2006). Effects of measurement method, wavelength and source-detector distance on the fast optical signal. Neuroimage32, 15761590. 10.1016/j.neuroimage.2006.05.030

  • 26

    GunningD.AhaD. (2019). DARPA’s explainable artificial intelligence (XAI) program. AI Mag.40, 4458. 10.1609/aimag.v40i2.2850

  • 27

    HochreiterS.SchmidhuberJ. (1997). Long short-term memory. Neural Comput.9, 17351780. 10.1162/neco.1997.9.8.1735

  • 28

    HolperL.WolfM. (2011). Single-trial classification of motor imagery differing in task complexity: a functional near-infrared spectroscopy study. J. Neuroeng. Rehabil.8:34. 10.1186/1743-0003-8-34

  • 29

    HomanR. W.HermanJ.PurdyP. (1987). Cerebral location of international 10-20 system electrode placement. Electroencephalogr. Clin. Neurophysiol.66, 376382. 10.1016/0013-4694(87)90206-9

  • 30

    HongK. S.NaseerN.KimY. H. (2015). Classification of prefrontal and motor cortex signals for three-class fNIRS-BCI. Neurosci. Lett.587, 8792. 10.1016/j.neulet.2014.12.029

  • 31

    HongK. S.SantosaH. (2016). Decoding four different sound-categories in the auditory cortex using functional near-infrared spectroscopy. Hear. Res.333, 157166. 10.1016/j.heares.2016.01.009

  • 32

    IraniF.PlatekS. M.BunceS.RuoccoA. C.ChuteD. (2007). Functional near infrared spectroscopy (fNIRS): an emerging neuroimaging technology with important applications for the study of brain disorders. Clin. Neuropsychol.21, 937. 10.1080/13854040600910018

  • 33

    JananiA.SasikalaM.ChhabraH.ShajilN.VenkatasubramanianG. (2020). Investigation of deep convolutional neural network for classification of motor imagery fNIRS signals for BCI applications. Biomed. Signal Process. Control62:102133. 10.1016/j.bspc.2020.102133

  • 34

    JoshiR.GoelP.SurM.MurthyH. A. (2018). “Single trial P300 classification using convolutional LSTM and deep learning ensembles method,” in Intelligent Human Computer Interaction, ed TiwaryU. (Cham: Springer), 315.

  • 35

    KhoaT. Q. D.NakagawaM. (2008). Functional near infrared spectroscope for cognition brain tasks by wavelets analysis and neural networks. Int. J. Psychol. Behav. Sci.2, 8994. 10.5281/zenodo.1082245

  • 36

    KingmaD. P.BaJ. (2017). Adam: a method for stochastic optimization. arXiv [Preprint]. 10.48550/arXiv.1412.6980

  • 37

    LiC.SuM.XuJ.JinH.SunL. (2020). A between-subject fNIRS-BCI study on detecting self-regulated intention during walking. IEEE Trans. Neural Syst. Rehabil. Eng.28, 531540. 10.1109/TNSRE.2020.2965628

  • 38

    LinW.TongT.GaoQ.GuoD.DuX.YangY.et al. (2018). Convolutional neural networks-based MRI image analysis for the Alzheimer’s disease prediction from mild cognitive impairment. Front. Neurosci.12:777. 10.3389/fnins.2018.00777

  • 39

    LiptonZ. C.BerkowitzJ.ElkanC. (2015). A critical review of recurrent neural networks for sequence learning. arXiv [Preprint]. 10.48550/arXiv.1506.00019

  • 40

    LuJ.YanH.ChangC.WangN. (2020). “Comparison of machine learning and deep learning approaches for decoding brain computer interface: an fNIRS study,” in Intelligent Information Processing X., eds ShiZ.VaderaS.ChangE. (Cham: Springer International Publishing), 192201.

  • 41

    LundbergS. M.LeeS. I. (2017). A unified approach to interpreting model predictions. Adv. Neural Inform. Process. Syst.30, 47654774. 10.48550/arXiv.1705.07874

  • 42

    LuoS. (2020). RTCOG: online ECoG-based brain-computer interface system for the decoding, synthesis and classification of speech neural signals. Available online at: http://jhir.library.jhu.edu/handle/1774.2/62726.

  • 43

    MinatiL.VisaniE.DowellN. G.MedfordN.CritchleyH. D. (2011). Variability comparison of simultaneous brain near-infrared spectroscopy (NIRS) and functional MRI (fMRI) during visual stimulation. J. Med. Eng. Technol.35, 370376. 10.3109/03091902.2011.595533

  • 44

    NagabushanamP.Thomas GeorgeS.RadhaS. (2020). EEG signal classification using LSTM and improved neural network algorithms. Soft Comput.24, 998110003. 10.1007/s00500-019-04515-0

  • 45

    NaseerN.HongK. S. (2013). Classification of functional near-infrared spectroscopy signals corresponding to the right- and left-wrist motor imagery for development of a brain-computer interface. Neurosci. Lett.553, 8489. 10.1016/j.neulet.2013.08.021

  • 46

    NaseerN.HongK. S. (2015). fNIRS-based brain-computer interfaces: a review. Front. Hum. Neurosci.9:3. 10.3389/fnhum.2015.00003

  • 47

    NaseerN.HongM. J.HongK. S. (2014). Online binary decision decoding using functional near-infrared spectroscopy for the development of brain-computer interface. Exp. Brain Res.232, 555564. 10.1007/s00221-013-3764-1

  • 48

    NaseerN.NooriF. M.QureshiN. K.HongK. S. (2016). Determining optimal feature-combination for LDA classification of functional near-infrared spectroscopy signals in brain-computer interface application. Front. Hum. Neurosci.10:237. 10.3389/fnhum.2016.00237

  • 49

    NazeerH.NaseerN.KhanR. A.NooriF. M.QureshiN. K.KhanU. S.et al. (2020). Enhancing classification accuracy of fNIRS-BCI using features acquired from vector-based phase analysis. J. Neural Eng.17:056025. 10.1088/1741-2552/abb417

  • 50

    NguyenA.KimJ.OhH.KimH.LinW.LeeS. (2019). Deep visual saliency on stereoscopic images. IEEE Trans. Image Process.28, 19391953. 10.1109/TIP.2018.2879408

  • 51

    O’SheaK.NashR. (2015). An introduction to convolutional neural networks. arXiv [Preprint]. 10.48550/arXiv.1511.08458

  • 52

    OlmosR.TabikS.LamasA.Pérez-HernándezF.HerreraF. (2019). A binocular image fusion approach for minimizing false positives in handgun detection with deep learning. Info. Fusion49, 271280. 10.1016/j.inffus.2018.11.015

  • 53

    OvchinnikovaA. O.VasilyevA. N.ZubarevI. P.KozyrskiyB. L.ShishkinS. L. (2021). MEG-based detection of voluntary eye fixations used to control a computer. Front. Neurosci.15:619591. 10.3389/fnins.2021.619591

  • 54

    PfeiferM. D.ScholkmannF.LabruyèreR. (2018). Signal processing in functional near-infrared spectroscopy (fNIRS): methodological differences lead to different statistical results. Front. Hum. Neurosci.11:641. 10.3389/fnhum.2017.00641

  • 55

    PowersD. M. (2020). Evaluation: from precision, recall and F-measure to ROC, informedness, markedness and correlation. arXiv [Preprint]. 10.48550/arXiv.2010.16061

  • 56

    RatheeD.RazaH.RoyS.PrasadG. (2021). A magnetoencephalography dataset for motor and cognitive imagery-based brain-computer interface. Sci. Data8:120. 10.1038/s41597-021-00899-7

  • 57

    RibeiroM. T.SinghS.GuestrinC. (2016). “Why should i trust you?,” in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (New York, NY, USA), 1135–1144. 10.1145/2939672.2939778

  • 58

    RiekeJ. D.MatarassoA. K.YusufaliM. M.RavindranA.AlcantaraJ.WhiteK. D.et al. (2020). Development of a combined, sequential real-time fMRI and fNIRS neurofeedback system to enhance motor learning after stroke. J. Neurosci. Methods341:108719. 10.1016/j.jneumeth.2020.108719

  • 59

    RoyS.YoussofzadehV.McCreadieK.PrasadG. (2020). Mapping BCI task imagery brain responses using MEG beta power desynchrony effects. Available online at: https://pure.ulster.ac.uk/en/publications/mapping-bci-task-imagery-brain-responses-using-meg-beta-power-des.

  • 60

    SantosaH.ZhaiX.FishburnF.HuppertT. (2018). The NIRS brain analyzIR toolbox. Algorithms11:73. 10.3390/a11050073

  • 61

    ShibuC. J.SreedharanS.KMA.KesavadasC. (2020). “Comparison of classification performance of handpicked, handcrafted and automated-features for fNIRS-BCI system,” in 2020 5th International Conference on Intelligent Informatics and Biomedical Sciences (ICIIBMS), (Okinawa, Japan), 152157. 10.1109/iciibms50712.2020.9336392

  • 62

    ShinJ.von LühmannA.BlankertzB.KimD. W.JeongJ.HwangH. J.et al. (2017). Open access dataset for EEG+NIRS single-trial classification. IEEE Trans. Neural Syst. Rehabil. Eng.25, 17351745. 10.1109/TNSRE.2016.2628057

  • 63

    ShrikumarA.GreensideP.ShcherbinaA.KundajeA. (2017). Not just a black box: learning important features through propagating activation differences. arXiv [Preprint]. 10.48550/arXiv.1605.01713

  • 64

    SimonJ.FishbeinP.ZhuL.RobertsM.MartinI. (2020). “Functional magnetic resonance imaging-based brain computer interfaces,” in Neural Interface Engineering: Linking the Physical World and the Nervous System, ed GuoL. (Cham: Springer International Publishing), 1747.

  • 65

    SitaramR.RosT.StoeckelL.HallerS.ScharnowskiF.Lewis-PeacockJ.et al. (2017). Closed-loop brain training: the science of neurofeedback. Nat. Rev. Neurosci.18, 86100. 10.1038/nrn.2016.164

  • 66

    SitaramR.ZhangH.GuanC.ThulasidasM.HoshiY.IshikawaA.et al. (2007). Temporal classification of multichannel near-infrared spectroscopy signals of motor imagery for developing a brain-computer interface. Neuroimage34, 14161427. 10.1016/j.neuroimage.2006.11.005

  • 67

    SorgerB.GoebelR. (2020). Real-time fMRI for brain-computer interfacing. Handb. Clin. Neurol.168, 289302. 10.1016/B978-0-444-63934-9.00021-4

  • 68

    StojicF.ChauT. (2020). Nonspecific visuospatial imagery as a novel mental task for online EEG-based BCI control. Int. J. Neur. Syst.30:2050026. 10.1142/S0129065720500264

  • 69

    SutskeverI.VinyalsO.LeQ. V. (2014). Sequence to sequence learning with neural networks. arXiv [Preprint]. 10.48550/arXiv.1409.3215

  • 70

    TaiK.ChauT. (2009). Single-trial classification of NIRS signals during emotional induction tasks: towards a corporeal machine interface. J. Neuroeng. Rehabil.6:39. 10.1186/1743-0003-6-39

  • 71

    TanveerM. A.KhanM. J.QureshiM. J.NaseerN.HongK. (2019). Enhanced drowsiness detection using deep learning: an fNIRS study. IEEE Access7, 137920137929. 10.1109/ACCESS.2019.2942838

  • 72

    TrakoolwilaiwanT.BehboodiB.LeeJ.KimK.ChoiJ. W. (2018). Convolutional neural network for high-accuracy functional near-infrared spectroscopy in a brain-computer interface: three-class classification of rest, right- and left-hand motor execution. Neurophotonics5:011008. 10.1117/1.NPh.5.1.011008

  • 73

    WangL.LiuR. (2020). Human activity recognition based on wearable sensor using hierarchical deep LSTM networks. Circuits Syst. Signal Process.39, 837856. 10.1007/s00034-019-01116-y

  • 74

    WolpawJ. R.BirbaumerN.McFarlandD. J.PfurtschellerG.VaughanT. M. (2002). Brain-computer interfaces for communication and control. Clin. Neurophysiol.113, 767791. 10.1016/s1388-2457(02)00057-3

  • 75

    YücelM. A.LühmannA. v.ScholkmannF.GervainJ.DanI.AyazH.et al. (2021). Best practices for fNIRS publications. Neurophotonics8:012101. 10.1117/1.NPh.8.1.012101

  • 76

    YingW.ZhangL.DengH. (2020). Sichuan dialect speech recognition with deep LSTM network. Front. Comput. Sci.14, 378387. 10.1007/s11704-018-8030-z

  • 77

    ZhangX.XiongQ.DaiY.XuX.SongG. (2020). An ECoG-based binary classification of BCI using optimized extreme learning machine. Complexity2020:e2913019. 10.1155/2020/2913019

  • 78

    ZhouX.SobczakG.McKayC. M.LitovskyR. Y. (2020). Comparing fNIRS signal qualities between approaches with and without short channels. PLoS One15:e0244186. 10.1371/journal.pone.0244186

Summary

Keywords

brain state classification, functional near-infrared spectroscopy, brain-computer interface, deep learning, convolutional neural networks, long short-term memory, explainable AI

Citation

Shibu CJ, Sreedharan S, Arun KM, Kesavadas C and Sitaram R (2023) Explainable artificial intelligence model to predict brain states from fNIRS signals. Front. Hum. Neurosci. 16:1029784. doi: 10.3389/fnhum.2022.1029784

Received

27 August 2022

Accepted

21 November 2022

Published

19 January 2023

Volume

16 - 2022

Edited by

Shenghong He, University of Oxford, United Kingdom

Reviewed by

Jinung An, Daegu Gyeongbuk Institute of Science and Technology (DGIST), South Korea; Xiaofeng Xie, Hainan University, China

Updates

Copyright

*Correspondence: Sujesh Sreedharan Ranganatha Sitaram

Specialty section: This article was submitted to Brain-Computer Interfaces, a section of the journal Frontiers in Human Neuroscience

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics