# Multi-Source Co-adaptation for EEG-Based Emotion Recognition by Mining Correlation Information

- Institute of Artificial Intelligence Application, Ningbo Polytechnic, Zhejiang, China

Since each individual subject may present completely different encephalogram (EEG) patterns with respect to other subjects, existing subject-independent emotion classifiers trained on data sampled from cross-subjects or cross-dataset generally fail to achieve sound accuracy. In this scenario, the domain adaptation technique could be employed to address this problem, which has recently got extensive attention due to its effectiveness on cross-distribution learning. Focusing on cross-subject or cross-dataset automated emotion recognition with EEG features, we propose in this article a robust multi-source co-adaptation framework by mining diverse correlation information (MACI) among domains and features with *l*_{2,1}−*norm* as well as correlation metric regularization. Specifically, by minimizing the statistical and semantic distribution differences between source and target domains, multiple subject-invariant classifiers can be learned together in a joint framework, which can make MACI use relevant knowledge from multiple sources by exploiting the developed correlation metric function. Comprehensive experimental evidence on DEAP and SEED datasets verifies the better performance of MACI in EEG-based emotion recognition.

## Introduction

Although emotion can be easily captured by human beings due to its close relationship with human’s cognition (Dolan, 2002), it cannot be readily recognized by instruments due to its complexity. Recently, as one of the most active research topics from the affective computing community, affection recognition had obtained a large amount of attention from pattern recognition and machine vision research fields (Kim et al., 2013). Generally, there are two categories on the responses of human emotion, i.e., external and internal responses. In this work, we focus on the latter. Basically, the representative internal responses include blood pressure, heart rate, respiration rate, electroencephalography (EEG), magneto encephalogram (Mühl et al., 2014), etc. Usually, the core components of a traditional emotion recognition system based on EEG are feature extraction and emotion classification (Lan et al., 2018). Practically, the time domain, frequency domain, and time–frequency domain are the main sources of the extracted EEG features (Jenke et al., 2014; Zhang et al., 2020b). The EEG feature extraction methods are more comprehensively reviewed in Jenke et al. (2014). In the past few years, aiming at the problem of emotion classification, a large number of emotion recognition models based on EEG signals have been proposed (Musha et al., 1997; Kim et al., 2013). For instance, a new group sparse canonical correlation analysis method was proposed in Zheng (2017) for simultaneous EEG channel selection and emotion recognition; in Li et al. (2018c), a graph regularized sparse linear regression method was proposed to deal with EEG-based emotion recognition. In the recent years, the deep learning method based on EEG has shown better performance than those traditional methods in emotion recognition and widely exploited in feature extraction and emotion recognition at the same time. For example, Zheng et al. (2015) employed deep belief network for EEG-based emotion recognition; Song et al. (2018) modeled the multi-channel EEG features by utilizing a graph and then performed EEG-based emotion recognition on these features; the work by Li et al. (2018b) had proposed a novel neural network model for EEG-based emotion recognition task.

While many models and methods for emotion recognition based on EEG have been proposed, most of them worked well only in the scenario that the training and test data were from the same distribution or domain. Under this hypothesis, the classifier trained on the source domain can directly predict the labels of the target data. However, for the problem of cross-domain emotion recognition based on EEG, many EEG-based emotion recognition methods would fail because of the distribution mismatch of EEG features. To this end, domain adaptation (DA) emotion recognition algorithms have emerged to investigate and address the automated emotion recognition problem (Chu et al., 2017), in which one has target domain with few or even none of labeled images by leveraging other related but different source/auxiliary domain(s) (Bruzzone and Marconcini, 2010). A typical example is the cross-subject EEG emotion recognition problem, in which the training and testing EEG data are from different subjects. To deal with the challenging cross-subject EEG emotion recognition problem, Pandey and Seeja (2019) proposed a subject-independent approach for EEG emotion recognition. Li et al. (2018a) proposed another method for cross-subject EEG emotion recognition. In the past decade, deep neural networks (Ganin et al., 2016; Li et al., 2018b) have also driven rapid progress in DA. The DA issues can be solved by the domain adversarial neural network (Ganin et al., 2016). It remains unclear, however, whether the performance of deep DA methods is really contributed by their deep feature representation, the fine-tuned classifiers, or is rather an outcome of the adaptation regularization terms (Ghifary et al., 2017).

Although the existing DA method has obvious effectiveness and efficiency in the special use of emotion recognition (Chu et al., 2017) in order to make use of the correlation knowledge among domains and features, there is little work to use the joint feature selection method and then carry out the multi-source adaptive domain recognition of cross-dataset. Besides this, during DA, most of the multi-source domain adaptation (MDA) methods (Yang et al., 2007; Tommasi et al., 2014) generally cope with the sources independently without considering the correlation information among the source domains, which may destroy the discriminant structure (either intrinsic or extrinsic) of multi-source domains. Last but not the least, for a MDA system, it is crucial for source weight determination during learning based on the correlation and quality of source domains. To the best of our knowledge, these characters are not feasible enough in extant MDA methods.

In order to solve the above-mentioned problems in existing MDA, we exploit the relevant information among sources and features to learn a multi-source emotion recognition model. We mainly adopt the strategy of digging the relationship between multi-source domains and one target domain (including feature and distribution) for promoting multi-source adaptive emotion recognition. We aim to progress beyond existing works that have partially addressed those issues by exploring to solve all the above-mentioned issues in a unified framework. Specifically, we develop a robust multi-source co-adaptation method for EEG-based emotion recognition by employing the correlation information (MACI) among features and sources *via l*_{2,1}−*norm* (Nie et al., 2010a) and correlation metric regularization. Under this framework, the correlation metric function is developed to mine the invariant knowledge among multi-source domains, the *l*_{2,1}−*norm* loss function aims to reduce the influence of outliers or noise, and row sparsity is designed to obtain the solution of sparse feature selection (Zhang et al., 2020b). We match distributions between each domain pair (including both target and multi-source domains) by minimizing the nonparametric maximum mean discrepancy (MMD) (Gretton et al., 2009; Pan et al., 2011) in a reproducing kernel Hilbert space (RKHS). The contributions of this article are listed as follows:

1. We propose a unified multi-source adaptive emotion recognition framework with EEG features by combining *l*_{2,1}−*norm* and correlation metric regularization.

2. Our framework selects features in a collaborative way and considers the correlated knowledge among features; the importance of each feature does not need to be evaluated separately. In addition, in our unified framework, we can learn multiple functions of feature selection for all source adaptation subjects synchronously so that our framework can use the correlated information of multiple sources as auxiliary information.

3. In this framework, the original geometric structure is retained by using the graph Laplacian regularization, and the *l*_{2,1}−*norm* minimization sparse regression approach is used to suppress the influence of noise or outliers in the domains, which shows the robustness of the framework.

4. Through a large number of experiments on two EEG datasets, we prove the effectiveness and convergence of this framework.

The remainder of the paper is organized as follows. In section “Related Work,” we discussed the related works with feature selection and MDA learning. In section “Proposed Framework,” our framework MACI will be designed, while section “Algorithm” arranges the corresponding optimal algorithm of MACI. Section “Algorithm Analysis” gives algorithm analysis, including the convergence and generalization. The experimental results and analysis on two real EEG datasets are presented in section “Experiments.” Finally, we conclude in section “Conclusion.”

## Related Work

In this section, we briefly review the prior emotion recognition with respect to EEG and multi-source adaptation techniques due to their relationships with our main ideas.

### Multi-Source Domain Adaptation

In the past several years, the mismatch problem between source and target domains has been solved by many DA technologies, which are widely used in a large number of visual applications, such as image annotation/classification, video concept detection, target recognition, and so on (Yang et al., 2007; Duan et al., 2012b, c; Tao et al., 2015, 2016, 2017; Ghifary et al., 2017). In the existing works for conquering DA, discovering one or multiple domain-invariant classifier(s) is a widely research topic (Yang et al., 2007; Tommasi et al., 2014) by constructing certain common subspace to make different sources have the same (or similar) marginal distributions. Therefore, the source classifier would obtain well performance on the target domain. Several methods have been studied to measure the distribution similarities (Gretton et al., 2009), and the analysis from them (Mansour et al., 2009) show that the performance of the classifier on the target is in positive correlation with those similar sources.

Very recently, to overcome the so-called negative transfer issue (Rosenstein et al., 2005), MDA methodology has been put forward by leveraging knowledge from multiple sources (Zhang et al., 2015, 2019c). A common problem in MDA is how to reduce the distribution difference between domains (Ghifary et al., 2017). To solve this issue, existing MDA approaches can be simply grouped into two classes (Tao et al., 2019), i.e., classifier-centric learning and feature-centric learning. The former is mainly based on the learning of the source domain classifiers in the target domain to adjust for realizing the implicit adaptation in the target distribution (e.g., Yang et al., 2007; Duan et al., 2012c; Tao et al., 2012; Tommasi et al., 2014), while the latter tries to accomplish the distribution alignment by learning a new representation of the data through a certain transformation (e.g., Wang and Mahadevan, 2011; Ghifary et al., 2017). This article focuses on the research of unsupervised classifier approaches.

In real application scenarios, the classifier-centric MDA scheme usually aims to directly design multiple adaptive source classifiers by merging the multiple distributions’ adaptation *via* feature representation or classifiers with model regularization. Lately, visual recognition works (Mansour et al., 2009; Nie et al., 2010a; Pan et al., 2011; Duan et al., 2012c; Tao et al., 2015, 2017) have proposed a great deal of classifier-centric MDA approaches. One part of classifier-centric MDA research assumes that there are enough number of unlabeled target instances and a large amount of labeled source instances in the training stage. Nevertheless, the remaining part of classifier-centric MDA research holds another hypothesis that only some labeled target instances are accessible in the training stage, which is also called model adaptation in the literature (Yang et al., 2007; Duan et al., 2012b, c; Tao et al., 2015, 2016, 2017; Ghifary et al., 2017). The model adaptation works effectively and efficiently just by exploiting the existing source models pre-trained on relevant but different source domains. Several representative state-of-the-arts include adaptive support vector machines (A-SVM) (Yang et al., 2007) *via* leveraging multiple source classifiers to suit a major target classifier, DA machine (or FastDAM) (Duan et al., 2012c) by employing sparsity regularizations and Laplacian manifold in least squares SVMs (Chai et al., 2016), etc. Recently, we also proposed some different model adaptation strategies (Tao et al., 2015, 2016, 2017) by leveraging the advantages of the low-rank and sparse representation.

### Emotion Recognition

In recent research about affective computing, increasing attentions have been paid on emotion recognition in the community of brain–computer interfaces (BCIs) (Mühl et al., 2014; Chu et al., 2017). An ideal emotion-based BCI can detect the emotional state through spontaneous EEG signals without explicit input from the user (Zhang et al., 2019b) and make a corresponding response to different emotional states. This kind of BCI may enhance the consumer experience in the time of an interactive session. Therefore, different approaches in Zhang et al. (2016, 2017) have been designed to recognize various emotion signals from brain wave. The latest affective BCIs (aBCIs) took machine learning algorithms and depended on a few features with discriminative information (Jenke et al., 2014; Mühl et al., 2014). When recording EEG signals in order to generate a desired target emotion signal, it is necessary to provide users with affective stimulation of specific emotions. In the training/calibration session, the required features and corresponding emotion labels are extracted from EEG signals to train the classifier. In an ongoing BCI session, the feature extractor receives the real-time EEG data and then sends the extracted features to the classifier for real-time affection classification. In this paradigm (Mühl et al., 2014), many researchers have reported pleasing classification performance. However, even if the experimental results are encouraging, the performance of aBCI still could be impacted by some reason. Since the EEG-based emotion signals are different from subject to subject, it is indispensable to train a specific object classifier for the subject of interest. Even in the same subject, the EEG signals are unstable, and the earlier trained classifier may perform poorly in the same subject at a later time. Therefore, in order to maintain a satisfactory classification accuracy, it is necessary to recalibrate frequently.

Domain adaptation method (Judy et al., 2017; Tzeng et al., 2017; Ding et al., 2018) has nearly completely dominated the recent literature of BCI (Jayaram et al., 2016). In aBCI studies, Dolan (2002), Koelstra et al. (2012), Shi et al. (2013), Mühl et al. (2014), Zheng et al. (2015), Zheng and Lu (2015, 2016), Chai et al. (2016, 2017), Lan et al. (2018), Zhong et al. (2020), and search various DA approaches by exploiting the SEED dataset.

## Proposed Framework

### Notations and Definitions

We describe the column vectors and matrices according to the small and capital letters, respectively, in this article. The often utilized symbols are listed in Table 1. The concatenation representation of *k* matrices according to row (horizontally) is like [*A*_{1}, *A*_{2}, …, *A*_{k}], and these matrices concatenation operations along a column (vertically) is denoted as [*A*_{1}; *A*_{2}; …; *A*_{k}]. The *l*_{2,1}−*norm* of *A* is defined as ${||A||}_{2,1}={\sum}_{i=1}^{n}{||{A}_{i,:}||}_{2}={\sum}_{i=1}^{n}\sqrt{{\sum}_{j=1}^{d}{A}_{ij}^{2}}$, and the trace-norm of *A* is indicated as ${||A||}_{*}=tr({(A{A}^{T})}^{{\scriptscriptstyle \frac{1}{2}}})$ (Lotfi and Akbarzadeh, 2014).

We mainly focus on the unsupervised MDA based on *S* various sources of *c*-class. Suppose there are *n*_{a} (*a* = 1, 2, …, *S*) instances in every source domain, respectively. In the *a*−th source domain, given ${X}^{a}=\{{x}_{1}^{a},\mathrm{\dots},{x}_{{n}_{a}}^{a}\}\in {\mathrm{R}}^{d\times {n}_{a}}\in \mathrm{\chi}$, and it is a training instances matrix with *c* sub-classes, which are associated with their class labels ${Y}^{a}={[{y}_{1}^{a},\mathrm{\dots},{y}_{{n}_{a}}^{a}]}^{T}\in {\mathrm{R}}^{{n}_{a}\times c}\in \mathrm{\Gamma}={\{0,1\}}^{c\times 1}$, a target domain dataset is denoted as ${X}^{t}=\{{x}_{1}^{t},{x}_{2}^{t},\mathrm{\dots},{x}_{m}^{t}\}\in {\mathrm{R}}^{d\times {n}_{t}}\in \mathrm{\chi}$, with their pseudo-class labels ${Y}^{t}={[{y}_{1}^{t},\mathrm{\dots},{y}_{{n}_{t}}^{t}]}^{T}\in {\mathrm{R}}^{{n}_{t}\times c}\in \mathrm{\Gamma}$ obtained from some supervised models (e.g., SVMs) which are trained on the source domain with labeled data. Our ultimate goal is to recognize the ground-truth class of test data ${x}_{k}^{t}\in {X}^{t}$, under the conditions that each domain pair *X ^{a}* and

*X*is assumed to be of different marginal and conditional distributions. While we do not need to limit that the instances number in each source domain is identical with that assumed when shaped into the training matrix, for the sake of simplicity, we can extract the same number of training instances from each source domain.

^{t}We further denote by ${X}^{a(\overline{l})}(\overline{l}=1,\mathrm{\dots},c)$ the set of samples in *X ^{a}* with the label $\overline{l}$. Similarly, the sample set in the target domain

*X*with the label $\overline{l}$ is defined as ${X}^{t(\overline{l})}$. Note that the true labels of the set ${X}^{t(\overline{l})}$ are unknown. We therefore employ in this work a base classifier, e.g., SVM, to attribute pseudo-labels for the subset in the target domain. For easy expression, we further define the matrix

^{t}*X*

_{a}= [

*X*,

^{a}*X*] ∈ R

^{t}^{d×N}(

*N*=

*n*

_{t}+

*n*

_{a}) with its label matrix

*Y*

_{a}= [

*Y*,

^{a}*Y*] in packing both source and target data with respect to the

^{t}*a*−th source domain.

** Definition 1 (MDA):** Let Δ = {P

^{1}, …, P

^{S}} be a set of

*S*source domains and P

^{t}∉Δ be a target domain. Denote by ${X}^{a}={\left\{{x}_{i}^{a},{y}_{i}^{a}\right\}}_{i=1}^{{n}_{a}}~{\text{P}}^{a}\left(a=1,\dots ,S\right)$ the samples drawn from the

*a*−th source domains and by ${X}^{t}={\left\{{x}_{i}^{t}\right\}}_{i=1}^{{n}_{t}}\sim {\mathrm{P}}^{t}$ the samples drawn from the target domain. The task of MDA is to learn an ensemble function ${f}_{{\text{P}}^{t}}:\mathcal{X}\to \mathrm{\Gamma}$ by co-learning multiple classifiers given

*X*(

^{a}*a*= 1, …,

*S*) and

*X*as the training examples.

^{t}** Definition 2 (Multi-source Domain Generalization):** In this scenario, the target domain is inaccessible in the training stage. Given

*S*source domains Δ = {P

^{1}, …, P

^{S}} and denoted by ${X}^{a}={\{{x}_{i}^{a},{y}_{i}^{a}\}}_{i=1}^{{n}_{a}}\sim {\mathrm{P}}^{a}$ the samples drawn from the

*a*−th source, the task of multi-source generalization is to co-learn multiple adaptive functions ${f}_{{\mathrm{P}}^{a}}:\mathcal{X}\to \mathrm{\Gamma}$ only given

*X*, ∀

^{a}*a*= 1, …,

*S*as the training examples, which could be well generalized to a certain unseen target domain.

### Problem Statement

In representative MDA, one can use the strategy of acquiring knowledge from multiple auxiliary sources to promote the target task of interest, which is better than learning each source task alone in emotion recognition. That is to say that common knowledge shared by multi-source domains is beneficial to emotion analysis. Moreover, some optimal recognition models have been developed in the latest works for the source domain and/or target domain separately. Furthermore, in these methods, joint multi-source adaption emotion recognition and feature selection has been largely unaddressed, and little or limited efforts have yet been devoted to the utilization of the correlated knowledge among sources.

To solve the above-mentioned issues, we propose in this work a robust multiple-source adaption emotion recognition method based on EEG features. The method utilizes the correlated knowledge among domains and features by joint *l*_{2,1}−*norm* and correlation metric regularization and can process high-dimensional, sparse, outliers, and non-i.i.d EEG data at the same time. The designed method has three characteristics, which are integrated into a unified optimization formulation to find an effective emotion recognition model and align the feature distribution between source and target domains: (1) *via* employing the *l*_{2,1}−*norm* minimization, a robust loss term is introduced to avoid the influence of noise or outliers in EEG signal, and a sparse regularization term is designed to eliminate over-fitting and a sparse feature subset is selected; (2) based on the designed regression model and the semantic distribution matching between each pair of domains, it not merely provides robustness on loss function but also retains the domain distribution (including local and global) structures and meanwhile maintains a high dependence on the (pseudo)-label knowledge of the source domains and the target domain (Zhang et al., 2020a) so as to obtain preferable generalization performance; and (3) through our constructed metric function of correlation, we can make full use of the correlative information among multiple sources and transfer more discriminative knowledge to the target domain. To implement these properties, in the following part, we will detail the objective formulation of the proposed method.

### General Formulation

In this section, we propose the general formulation of MACI framework underpinned by the robust regression principle and the regularization theory. In particular, our main purpose is to optimize a unified objective function by compromising the following three complementary objectives:

1. Robust multi-source co-regression with feature selection using *l*_{2,1}−*norm* minimization, in which the domain label consistency is explicitly maximized through iterative linear label regression.

2. Aligning domain distributions including global statistical distributions and intra-domain semantic distributions or class conditional distributions.

3. Effectively utilizing correlation information among source domains *via* developing an effective correlation metric function.

For the multi-source adaptation emotion recognition of interest, we define the *a*−th (*a* = 1, …, *S*) classifier function as ${f}_{a}({X}_{a})={X}_{a}^{T}{W}_{a}$, where *W*_{a} is the *a*−th classifier model, and *W*_{0} is certain reference model. Suppose there is a kernel feature map ϕ_{a}: χ→*H*_{a} ^{1} that projects the training data from the original feature space into certain RKHS (Nie et al., 2010b) *H*_{a}, the predictor weight *W*_{a} can be kernelized. We denote the kernel matrix as ${\left({K}_{a}\right)}_{i,j}=\u27e8\mathrm{\varphi}({x}_{i}^{a}),\mathrm{\varphi}({x}_{j}^{a})\u27e9$, where ${x}_{i}^{a},{x}_{j}^{a}\in {X}_{a}$. We present the empirical kernel map as discussed in Gretton et al. (2009):

We therefore have kernel matrices *K*_{a} = ϕ_{a}(*X*_{a}). Hence, the kernelized decision function on *X*_{a} (*a* = 1, …, *S*) becomes ${f}_{a}({X}_{a})={K}_{a}^{T}{W}_{a}$. We further denote by *W* = [*W*_{1}; …; *W*_{S}] the concatenation matrix.

We then endeavor to find *S* cross-domain models parameterized by ${\{{W}_{a}\}}_{a=1}^{S}$ in some empirical RKHSs *via* jointly utilizing correlated knowledge among sources and features. In view of the above-cited objectives, we propose the following general formulation of MACI.

where *R*(⋅,⋅) is the robust regression function with feature selection *via l*_{2,1}−*norm* minimization, Ω_{A}(*X ^{a}*,

*X*) is certain distance metric function for aligning the domain distributions, and

^{t}*Cor*(⋅) is a correlation metric function which is a global regularization term. In the subsequent sections, we focus on designing these components in the general formulation one by one to construct a unified framework.

### Design of Robust Multi-Source Co-regression With Feature Selection

To achieve the first objective mentioned above, one should jointly minimize each source regression loss and implement feature selection, in which the domain label consistency is explicitly maximized, and the data outliers are accounted for to avoid negative transfer. To this end, we first explain a predicted label matrix *F*_{a} ∈ R^{N×c}(*a* = 1, …, *S*) into our predictive function (Nie et al., 2010b). The predicted values in this label matrix should satisfy local smoothness and global consistency, i.e., they should preserve the local geometry while fitting in with the true labels (Zhang et al., 2020a). To satisfy these requirements, we present a smooth regularization term on the label geometric structure between each source instance (Nie et al., 2010b; Yan et al., 2006), which is formulated as

where Ů is a diagonal matrix with Ů_{i,i} = *ς* (_{ς} is a large specified value) if ${x}_{i}^{a}\in {X}_{a}$ has a label, Ů_{i,i} = 0 or else and *a* is a regularization parameter. *L*_{a} is the graph Laplacian matrix of the *a*−th source dataset, which is defined as *L*_{a} = Λ_{a}−∏_{a}, where Λ_{a} is a diagonal matrix with (Λ_{a})_{i,i} = ∑_{j}(∏_{a})_{i,j}, and ∏_{a} is the weight matrix of the graph, which can be defined as:

where the * _{k}* nearest neighbors of

*x*are assigned to δ

_{k}(

*x*), and γ

_{a}is a hyper-parameter, which can be empirically selected as ${\overline{\mathrm{\theta}}}_{a}\sqrt{c}$ by considering the impact of multi-class distribution on the affinity relationship among the domain data, where ${\overline{\mathrm{\theta}}}_{a}$ is the square root of the mean norm of

_{Xa}.

We therefore design the following multi-source sparse co-regression model for meeting the first objective.

where ϑ = [ϑ_{1}, …, ϑ_{a}]^{T} is the weight vector to jointly combine all source regression loss, β is a regularization parameter, and *q*_{1} > 1 is a tunable parameter for avoiding trivial solution. The model (2) is convex, and the *l*_{2,1}−*norm* loss function ${||{K}_{a}^{T}{W}_{a}-{F}_{a}||}_{2,1}$ is robust to outliers (Li et al., 2015). In the meantime, the term ||*W*_{a}||_{2,1} assures that *W*_{a} can accomplish feature selection across different domains due to its sparsity. That is, by exploiting the correlation among different features, our approach can jointly evaluate all feature knowledge of source domains and target domain.

#### Design of Domain Distribution Alignment

As a nonparametric distribution discrepancy estimator, MMD (Gretton et al., 2009) was used to compare two distributions by transforming the distributions into a RKHS (Pan et al., 2011; Duan et al., 2012b; Tao et al., 2012; Chen et al., 2013; Long et al., 2014). Let ℱ be a set of functions *f*: 𝒳 → R. The MMD between two domains P and Q is defined as

The MMD measures the similarity level between two domains from the side of function class ℱ. To make the MMD a proper regularization for the classifier model *W*_{a}, we adopt the following the empirical estimation of MMD between *X ^{a}* and

*X*, which is defined as

^{t}where ||⋅||_{H} is the RKHS norm, ${({K}_{a})}_{i,j}=<{\mathrm{\varphi}}_{a}({x}_{i}^{a}),{\mathrm{\varphi}}_{a}({x}_{j}^{a})>$ with ${x}_{i}^{a},{x}_{j}^{a}\in {X}_{a}$, and

As for *MMD*_{e}(*X ^{a}*,

*X*) in Eq. 4, whereas even if there is a perfect domain distribution match, it does not assure the instances from different domains, but the same class of labels will be mapped near the transform space. Lack of semantic consistency will be a major reason for performance degradation. Therefore, we use the following terms to develop a semantically matched MMD (Long et al., 2014):

^{t}where ${K}_{a(l)}={\mathrm{\varphi}}_{a}([{X}^{a(\overline{l})},{X}^{t(\overline{l})}])$ with ${({K}_{a(l)})}_{i,j}=<{\mathrm{\varphi}}_{a}({x}_{i}^{a(\overline{l})}),{\mathrm{\varphi}}_{a}({x}_{j}^{a(\overline{l})})>$, ${x}_{i}^{a(\overline{l})}\in {X}^{a(\overline{l})}$ and ${x}_{j}^{t(\overline{l})}\in {X}^{t(\overline{l})}$, and

We call Eq. 7 conditional (or semantic) MMD, which explicitly encourages instances from various domains but with the same label to map to the nearest in multi-source subspace. Finally, we suggest that the domain distribution alignment could be approached by learning multiple optimal models such that

where *C*_{a(0)} = *K*_{a}*M*_{a}*K*_{a} and ${C}_{a(\overline{l})}={K}_{a(\overline{l})}{M}_{a(\overline{l})}{K}_{a(\overline{l})}(\overline{l}=1,\mathrm{\dots},c)$. Let ${C}_{a}={\sum}_{\overline{l}=0}^{c}{C}_{a(\overline{l})}$, then we have ${\mathrm{\Omega}}_{A}({X}^{a},{X}^{t})=tr({W}_{a}^{T}{C}_{a}{W}_{a})$.

#### Design of Correlation Metric Function

As we know, a commonly used strategy in extant classifier-centric adaptation methods (Duan et al., 2012c; Tommasi et al., 2014) is to directly match the discriminant models between different domains, which is defined as:

** Definition 3 (model discriminant discrepancy, MDD):** Let 𝒲 be a set of function parameters 𝒲: 𝒳 → R. The model discriminant discrepancy between domains P and Q is defined as

It may be difficult to push these two models respectively learnt from different domains when the distribution discrepancy between them is large. In our correlation metric function, we instead aim to guarantee each source model to be aligned with a global reference matrix *W*_{0} so as to enable different source models to share the common knowledge for effectively utilizing correlation information among source domains. In essence, *W*_{0} builds a transformation among source domains so that knowledge of one source can be used to another. They yield the following model alignment function (MAF):

** Definition 4 (model alignment function):** Given

*S*domains ${\{{X}^{a}\}}_{a=1}^{S}$ on 𝒳, we can think of the classification model set ${\{{W}^{a}\}}_{a=1}^{S}$ in some latent spaces. Their MAF is then defined as

where η = [η_{1}, …, η_{S}]^{T} is a weight vector for discriminatively selecting different source knowledge with ${\sum}_{a=1}^{S}{\mathrm{\eta}}_{a}=1$, and *W*_{0} is certain shared (common) discriminant model among these domains (Zhang et al., 2019a).

In essence, the MAF measures the similarity between two domain classifiers by the classification model. The next theorem is about from MAF to MDD between two domains.

** Theorem 1 (MAF bounds MDD):** The (squared) maximum discriminant discrepancy between domains P and Q on 𝒳 is upper-bounded by their MAF with η

_{P}= η

_{Q}= η

_{Δ}:

where 𝒲 = {*W*: 𝒳 → R|*W* is the classifier model} and *W*_{P}, *W*_{Q} ∈ 𝒲. Specially if 𝒲 is induced by a characteristic kernel on 𝒳, then Ψ ({*W*_{P}, *W*_{Q}}) = 0 if and only if P = Q.

**Proof.** By definition 4 and triangle inequality theorem,

that is, Ψ({*W*_{P}, *W*_{Q}}) bounds $MD{D}_{\mathcal{W}}^{2}[\mathrm{P},\mathrm{Q}]$. If 𝒲 is induced by some characteristic kernel on 𝒳, then Ψ({*W*_{P}, *W*_{Q}}) = 0 if and only if P = Q, which can be concluded from the result of Theorem 2.2 of Gretton et al. (2009).

Theorem 1 also indicates that the MAF is an effective metric method if the kernel on 𝒳 is characteristic (Gretton et al., 2009).

** Theorem 2:** In particular, if

*W*

_{0}= η

_{P}

*W*

_{P}+ η

_{Q}

*W*

_{Q}in MAF, we obtain

**Proof.** By η_{P} + η_{Q} = 1, we have

where the last inequality follows after observing that η_{P}η_{Q} ≤ 1/4 with the equality holding when η_{P} = η_{Q} = 1/2.

For achieving the third target of ours, the correlation metric function therefore was designed as follows:

where *q*_{2} > 1 is another tunable parameter to avoid a trivial solution. The regularization term ||*W*||_{*} in Eq. 9 enables different projection functions ${\{{W}_{a}\}}_{a=1}^{S}$ to share common information/parts through models of sources. Thus, the knowledge from multiple sources can be further shifted from one by one source domain.

### Final Formulation

By using empirical kernel function and integrating Eqs 2, 8, and 9, we therefore propose the following unified framework to implement the combination of feature selection and domain adaptive learning by utilizing the correlated knowledge across multi-sources and features.

Note that, in Eq. 10, the *l*_{2,1}−*norm* loss function makes the model effectively robust to noises or outliers from domains. In addition, after *l*_{2,1}−*norm* regularization is added to *W*_{a}, many rows in *W*_{a}(*a* = 1, …, *S*) become zero. Therefore, the characteristics corresponding to these rows with zeros are not significant for the target task learning. Therefore, for acquiring competitive performance, we can select features from the original domain data. Similarly (Nie et al., 2010a), we sort the rows in *W*_{a} by descending sequence in light of each row of values in *l*_{2}−*norm* and followed by selecting the top rows as the feature selection outcome.

** Remark 1:** Our proposed method (10) has some competitions in universalization and efficiency. Firstly, in order to deal with the change of feature dimensions and types in among source domains easily, we learn a separate classification model for each independent domain pair (i.e., one source domain and one target domain). Then, for dealing with more heterogeneous sources, the algorithm is easy to extend. Moreover, in order to improve the speed of the algorithm, the redundant and irrelevant knowledge in the original features is thrown away before classification through sparse feature selection method, and then the classification models are learned. Furthermore, joint

*l*

_{2,1}−

*norm*and trace-norm minimization can be used to express well the together learning of feature selection and classification models so as to assure that the common subspace in the multi-domain can be extracted.

## Algorithm

In this section, we first give an iterative approach to optimize the objective function (10) followed by its complexity and classification function. Although matrix completion is realized by an alike optimization method from Yang et al. (2013), we focus on the other issue, that is, joint optimization of trace norm and *l*_{2,1}−*norm*. We then further present an effective and valuable extension to domain generalization when the target domain is inaccessible.

### Algorithm Optimization

According to Nie et al. (2010a), the derivative of *tr*(T̊^{T}*Q*T̊) is equal to the derivative of ||T̊||_{2,1}, i.e., 2*tr*(T̊^{T}*Q*T̊) = ||T̊||_{2,1}, where *Q* is a diagonal matrix and its *i*−th diagonal value is ${Q}_{ii}=\frac{1}{2{||{\text{T\u030a}}_{i,:}||}_{2}}$, and if T̊_{i,:} = 0, we can let ${Q}_{ii}=\frac{1}{2{||{\text{T\u030a}}_{i,:}||}_{2}+\mathrm{\epsilon}}$, where ε is a very small given value. Hence, we can farther transform Eq. 10 into Eq. 11:

where ${S}_{a}={[\underset{a-1}{\underset{\u23df}{{0}_{r},\mathrm{\dots},{0}_{r}}},{I}_{r},\underset{S-a}{\underset{\u23df}{{0}_{r},\mathrm{\dots},{0}_{r}}}]}^{T}$, the diagonal matrix *G*_{a} is based on *W*_{a}, where the *k*−th element is equal to ${({G}_{a})}_{kk}=\frac{1}{2{||{({W}_{a})}_{k,:}||}_{2}}$, the diagonal matrix *Z*_{a} is based on ${T}_{a}=\sqrt{{\mathrm{\vartheta}}_{a}^{{q}_{1}}}({K}_{a}^{T}{W}_{a}-{F}_{a})$, and where the *k*−th entry is computed by ${({Z}_{a})}_{kk}=\frac{1}{2{||{({T}_{a})}_{k,:}||}_{2}}$. By taking the derivative of Eq. (11) in reference to *W*_{0} and equaling to zero, we obtain:

where ${\mathrm{\eta}}_{0}=[{\stackrel{~}{\mathrm{\eta}}}_{1}{I}_{r};\mathrm{\dots};{\stackrel{~}{\mathrm{\eta}}}_{S}{I}_{r}]$ with ${\stackrel{~}{\mathrm{\eta}}}_{a}={\mathrm{\eta}}_{a}^{{q}_{2}}\mathrm{/}{\sum}_{a=1}^{S}{\mathrm{\eta}}_{a}^{{q}_{2}}(a=1,\mathrm{\dots},S)$. Substituting Eq. 12 into Eq. 11, we have

Note that the sub-gradient matrices *Z*_{a} and *G*_{a} in Eq. 13 are dependent on the matrices *W*_{a} and *F*_{a}, which are also unknown beforehand. Consequently, the objective function in Eq. 13 is a multi-variable optimization problem involving the variables *W*_{a}, *F*_{a}, ϑ_{a}, η_{a}. Since optimizing these variables simultaneously is difficult, we exploit the alternating iterative strategy to update one variable iteratively while the other variable(s) is(are) fixed. Therefore, the problem in Eq. 13 can be decomposed into four sets of convex sub-problems. We aim to find the optimal solution to each sub-problem alternatively and iteratively so that the objective function in Eq. 13 would converge to a local optimal solution. By initializing *W*_{a} and *F*_{a}, thus initializing *Z*_{a} and *G*_{a}, then we can start the iterations.

#### Optimize *W*_{a} and *F*_{a} by Fixing ϑ and η

By solving the derivative of Eq. 13 in reference to *W*_{a} and equaling to zero, we obtain:

where

and

Plugging Eq. 14 into Eq. 13, by mathematical calculating, we can get:

where

Lastly, substituting the optimal solution of the other variables into Eq. 18 to update *F*_{a}, the constraints ${F}_{a}^{T}{F}_{a}={I}_{c}$ should be added additionally, and (*F*_{a})_{ij} ≥ 0. Then, we can get the objective function in reference to *F*_{a}:

where ζ ^{2} and θ are balance arguments. By solving the derivative of Eq. 19 in reference to *F*_{i,j} and equaling to zero and exploiting the K.K.T. with constraint term θ_{ij}*F*_{i,j} = 0, we can get:

#### Optimize ϑ_{a} by Fixing *W*_{a}, *F*_{a}, and η_{a}

In this situation, the issue in Eq. 13 changes to a small problem as follows:

Let ${g}_{a}=tr\left({T}_{a}^{T}{Z}_{a}{T}_{a}\right)+tr({W}_{a}^{T}{C}_{a}{W}_{a})$, the Lagrange function of Eq. 21 is

Setting the derivative of ℑ(ϑ_{a}, ϕ) with respect to ϑ_{a} is equivalent to 0, and we can obtain:

Substituting Eq. 23 into the constraint ${\sum}_{a=1}^{S}{\mathrm{\vartheta}}_{a}=1$, we obtain

#### Optimize η_{a} by Fixing *W*_{a}, *F*_{a}, and ϑ_{a}

By fixing *W*_{a}, *F*_{a}, and ϑ_{a}, the problem in Eq. 13 then becomes the following sub-problem:

Let ${h}_{a}={||W{S}_{a}-W{\mathrm{\eta}}_{0}||}_{F}^{2}$, the Lagrange function of Eq. 25 is

Setting the derivative of ℑ(η_{a}, ρ) in reference to η_{a} is equivalent to 0, and we then get:

Substituting Eq. 27 into the constraint ${\sum}_{a=1}^{S}{\mathrm{\eta}}_{a}=1$, we obtain:

#### Overall Procedure

In this sub-section, we finally report the whole optimization process of MACI in Algorithm 1, where a window-based breaking criterion is employed to better obtain the convergence of the algorithm (Zhang et al., 2019a). Concretely speaking, defining a window size ℏ, we compute *ς* = ||*Max*Θ_{itr}−*Min*Θ_{itr}||/*Max*Θ_{itr} in *itr*−th iteration, where Θ_{itr} = {*Obj*_{itr−ℏ + 1}, …, *Obj*_{itr}} represents the set of historical target values in the window. When *ς* is less than a given threshold ε, that is *ς* < ε, the algorithm stops iterating. In our experiments, we set ε = 10^{−5} and ℏ = 6 empirically without losing statistical performance. We will discuss in section “Convergence” why Algorithm 1 is convergent.

### Computational Complexity

In this subsection, we give a formal analysis about the computational complexity of several main components in Algorithm 1 using the big *O* notation. Firstly, the construction of the *k*−*NN* graph and computing of the kernel matrix ${\{{K}_{i}\}}_{i=1}^{S}$, respectively, need computational cost $O(Sd{n}_{a}^{2})$ and *O*(*SdN*^{2}). Then, the optimization proceeds according to step by step iteratively. The cost for computing *F*_{a} is $O\left(3{n}_{a}^{3}+{n}_{a}^{2}c\right)$. After *F*_{a} was updated, computing *W*_{a} would cost *O*(*n*_{a}*c*^{2} + *dc*^{2}). In a word, the whole calculating cost is $O\left(\mathrm{\ell}S\left(3{n}_{a}^{3}+{n}_{a}^{2}c+{n}_{a}{c}^{2}+d{c}^{2}\right)+Sd{n}_{a}^{2}+Sd{N}^{2}\right)$. We assume that all Laplacian matrixes ${\{{L}_{i}\}}_{i=1}^{S}$ can be pre-calculated before the iterative optimization of Algorithm 1, and multi-kernel can be pre-calculated and put into memory before training. Thus, Algorithm 1 is effective and efficient computationally.

### Target Classification

The datasets of the unlabeled samples from the target domain are defined as ${K}_{u}^{t}={\{({k}_{j}^{t})\}}_{j=1}^{{n}_{t}}$. In the unsupervised DA learning case, one can predict a target class using the classification model *W*_{0}. Specifically, one may use $\mathrm{arg}\underset{1\le j\le c}{max}{({\mathrm{\Gamma}}_{u}^{t})}_{j}$ to classify a test sample ${k}_{u}^{t}\in {K}_{u}^{t}$ into one of the *c* target classes, where ${\mathrm{\Gamma}}_{u}^{t}={({W}_{0})}^{T}{k}_{u}^{t}$.

In our multi-source adaptation framework, nevertheless, another voting method defined as “sum” can be deduced, that is, once ${\{{W}_{a}^{s}\}}_{a=1}^{S}$ are obtained, for a test data ${k}_{u}^{t}\in {K}_{u}^{t}$, we can learn its label vector ${\mathrm{\Gamma}}_{u}^{t}$ by minimizing the residue between ${\mathrm{\Gamma}}_{u}^{t}$ and the projected vector of each source model:

The result of Eq. 29 can be acquired according to the constraint term ${\sum}_{a=1}^{S}{\mathrm{\vartheta}}_{a}=1$:

Once ${\mathrm{\Gamma}}_{u}^{t}$ is computed by using Eq. 30, we then use $\mathrm{arg}\underset{1\le j\le c}{max}{({\mathrm{\Gamma}}_{u}^{t})}_{j}$ to determine the class for this test data.

## Algorithm Analysis

### Convergence

We start with the next two lemmas and then demonstrate that the alternant optimization process, namely, step 2 in Algorithm 1, in the optimization issue of the Eq. 10, the optimal solution of ${\{{W}_{i}\}}_{i=1}^{S}$ converges.

** Lemma 1.** (Nie et al., 2010a) There are any two values

*V*

_{1},

*V*

_{2}∈ R

^{d}, and they are not equal to zero; we can get the inequality as:

** Lemma 2.** (Nie et al., 2010a) For any invertible matrices P̊ and Q̊, the following inequality holds:

Then, the iterative method designed in Algorithm 1 can converge to the optimal solution, which will be proved in the next theorem.

** Theorem 3:** In each iteration of Algorithm 1, the objective function of issue in Eq. 10 will be monotonically decreasing and finally will converge to the optimal solution of the issue.

**Proof.** For easy description, we define the updated *W*_{i} and *F*_{i} in the iteration τ as ${W}_{i}^{\mathrm{\tau}}$ and ${F}_{i}^{\mathrm{\tau}}(i=1,\mathrm{\dots},S)$ separately. The updating from step 2 of Algorithm 1 is equivalent to the optimum of the next problem:

Following the expressions of *Z*_{i}, *G*_{i}, and *U*, then we can get:

We can have the next inequality by Lemma 1:

Therefore, we have

Eq. 35 can be further rewritten as:

Noting that ${U}^{l}=\frac{1}{2}{\left({W}^{\mathrm{\tau}}{\left({W}^{\mathrm{\tau}}\right)}^{T}\right)}^{-{\scriptscriptstyle \frac{1}{2}}}$ and according to Lemma 2, we get

Subtracting Eq. (37) from Eq. 36, we have

That is to say

Hence, the theorem has been verified.

According to the optimization strategy of Algorithm 1, the objective function is monotonically decreasing in problem Eq. 10, so it is easiest to observe that the algorithm is convergent.

### Generalization

In this part, we derive an empirical bound for our method that shows how both MAF and PDS control the generalization performance under the situation of the squared loss *loss*(*a*, *b*) = (*a*−*b*)^{2}. The main idea is to merge the domain scatter into the proven adaptive range for the distance difference (Ghifary et al., 2017).

Denote by *H*:={*h*: 𝒳 → 𝒴} a hypothesis class of functions in the RKHS ℋ, where 𝒳 is a compact set and 𝒴 is a label space. Given a loss function *loss*(⋅,⋅): 𝒴 × 𝒴 → R_{+} and a domain distribution 𝒟 over 𝒳, we denote by ℒ_{𝒟}(*h*, h̊) = *E*_{x∼𝒟}[*loss*(*h*(*x*), h̊(*x*))] the expected loss for the given two functions *h*, h̊ ∈ *H*. Then, the distance of domain difference between two distributions P and Q is defined as:

By the notation in Eq. 40, we can obtain domain generalization bounds by domain scatter. Let *f*_{P} and *f*_{Q} be the true labeling functions for domain P and Q, respectively, and ${h}_{\mathrm{P}}^{*}:=\underset{h\in H}{\text{argmin}}{\mathcal{L}}_{\mathrm{P}}(h,{f}_{\mathrm{P}})$ and ${h}_{\mathrm{Q}}^{*}:=\underset{h\in H}{\text{argmin}}{\mathcal{L}}_{\mathrm{Q}}(h,{f}_{\mathrm{Q}})$ be the minimizers. The following theorem provides adaptation bounds with PDS (….).

** Theorem 4 (adaptation bounds with PDS)** (Ghifary et al., 2017): Denote by

*H*:={

*f*∈ ℋ: 𝒳 → R, ||

*f*||

_{ℋ}≤ 1

*and*||

*f*||

_{∞}≤

*r*} a class of functions in the RKHS ℋ and by ${X}_{\mathcal{X}}^{\mathrm{P}}=({x}_{1}^{s},\mathrm{\dots},{x}_{{n}_{s}}^{s})\sim \mathrm{P}$ and ${X}_{\mathcal{X}}^{\mathrm{Q}}=({x}_{1}^{t},\mathrm{\dots},{x}_{{n}_{t}}^{t})\sim \mathrm{Q}$ the source and target dataset, respectively. Let

*loss*(⋅,⋅): 𝒴 × 𝒴 → [0, Υ] be a

*q*−Lipschitz loss function, i.e., for all

*a*,

*b*∈ 𝒴 × 𝒴, ||

*loss*(

*a*)−

*loss*(

*b*)|| =

*q*||

*a*−

*b*||. Then, for any hypothesis

*h*∈

*H*, with probability of at least 1 − δ, the following generalization bound holds with the Rademacher complexity ${\mathrm{\Re}}_{{X}_{\mathcal{X}}^{\mathrm{P}}}(H)$ over ${X}_{\mathcal{X}}^{\mathrm{P}}$:

Theorem 4 provides a generalization bound for DA by introducing PDS and Rademacher complexity that measures the level to which a class of functions can fit random noise. The Rademacher complexity measure is the basis of relating empirical loss with expected loss. From Theorem 4, for a successful DA, we shall make ${\mathcal{L}}_{\mathrm{P}}({h}_{\mathrm{P}}^{*},{h}_{\mathrm{Q}}^{*})$ as small as possible. According to definition 3, the (squared) loss ${\mathcal{L}}_{\mathrm{P}}({h}_{\mathrm{P}}^{*},{h}_{\mathrm{Q}}^{*})$ is essentially equivalent to MDD in some optimal RKHS. We then further provide the following adaptation bounds with PDS and MAF, which follows by Theorem 1 combined with Theorem 4.

** Theorem 5 (adaptation bounds with PDS and MAF):** Denote by

*H*:={

*f*∈ ℋ: 𝒳 → R, ||

*f*||

_{ℋ}≤ 1 and ||

*f*||

_{∞}≤

*r*} a class of functions in the RKHS ℋ and by ${X}_{\mathcal{X}}^{\mathrm{P}}=({x}_{1}^{s},\mathrm{\dots},{x}_{{n}_{s}}^{s})\sim \mathrm{P}$ and ${X}_{\mathcal{X}}^{\mathrm{Q}}=({x}_{1}^{t},\mathrm{\dots},{x}_{{n}_{t}}^{t})\sim \mathrm{Q}$ the source and target dataset, respectively. Let

*loss*(⋅,⋅):𝒴 × 𝒴 →[0,Υ] be a

*q*−Lipschitz loss function, i.e., for all

*a*,

*b*∈ 𝒴 × 𝒴, ||

*loss*(

*a*)−

*loss*(

*b*)|| =

*q*||

*a*−

*b*||. Denote by ${W}_{\mathrm{P}}^{*},{W}_{\mathrm{Q}}^{*}$ the optimal functions learnt from domain P and domain Q, respectively; then, for any hypothesis

*h*∈

*H*, with probability of at least 1−δ, the following generalization bound holds with the Rademacher complexity ${\mathrm{\Re}}_{{X}_{\mathcal{X}}^{\mathrm{P}}}(H)$ over ${X}_{\mathcal{X}}^{\mathrm{P}}$ :

Theorem 5 clearly shows that the projected domain scatter Ξ_{ϕ}({μ_{Q}, μ_{P}}) and MAF $\mathrm{\Psi}\left(\{{W}_{\mathrm{P}}^{*},{W}_{\mathrm{Q}}^{*}\}\right)$ can control the generalization performance of MACI with its empirical measure, that is, to minimize the PDS (or, alternatively, the distributional scatter discrepancy) and MAF (or model discrimination discrepancy) in our methods can effectively improve the generalization bound in the setting of MDA or domain generalization, which is also supported by the following real-world experiments.

## Experiments

In this section, to evaluate the effectiveness of MACI for emotion recognition, we compare it with several state-of-the-art methods on two benchmark datasets, i.e., DEAP (Koelstra et al., 2012) and SEED (Zheng and Lu, 2015), which are also widely adopted as benchmark datasets for EEG-based emotion recognition (Mansour et al., 2009). Since existing deep DA models have demonstrated to be very effective, mainly applied to the EEG-based emotion recognition problems (Lotfi and Akbarzadeh, 2014), we divide our experiments into two parts, i.e., comparisons with shallow (traditional) DA methods on those emotion recognition tasks mentioned above and comparisons with the deep (CNN-based) DA methods for EEG-based emotion recognition on several cross-datasets.

## Data Preparation

At present, there are some EEG datasets for emotional state research. In this article, we used the following two public datasets: DEAP (Koelstra et al., 2012) and SEED (Zheng and Lu, 2015). As reported in Zhong et al. (2020) and Lan et al. (2018), there is a significant difference between these two databases due to some technical aspects. We also adopted the same feature extraction strategy with that in Lan et al. (2018). More details about these two databases can be found in Lan et al. (2018).

In our experiments, differential entropy (DE) (Zhong et al., 2020) is employed as the feature of emotion recognition. In the literature (Shi et al., 2013; Zheng et al., 2015; Chai et al., 2016; Zheng and Lu, 2016; Chai et al., 2017; Lan et al., 2018; Zhong et al., 2020) about the DA emotion recognition based on EEG, DE features have been widely used. The details of DE are explained in Lan et al. (2018).

### Baseline Setting

We compare our MACI method with the following state-of-the-art (related) baselines for multi-source emotion recognition tasks. Besides this, we also report the emotion recognition results of MACI using several deep features:

• No adaptation baseline FSSL (Yang et al., 2013)

• Multi-kernel adaptation method: FastDAM (Duan et al., 2012c)^{3}

• Multi-KT (Tommasi et al., 2014)^{4} : according to Tommasi et al. (2014), we here also use the *l*_{2}−*norm* constraint on *p* in Multi-KT algorithm

• Adaptive SVM: A-SVM (Yang et al., 2007)^{5}

• Domain selection machine (DSM) (Duan et al., 2012a)

• Deep DA methods: DAN (Long et al., 2015) and ReverseGrad (Ganin and Lempitsky, 2015).

For the baseline FSSL without adaptation and the multi-source adaptation method A-SVM, we just equally fuse the decision values of all base classifiers with each classifier learned on one source domain^{6}.

In our MACI, only several vital parameters such as *q*_{1}, *q*_{2}, λ, α, and β in our model need to be predefined. Considering that parameter determination is a yet unaddressed open issue, we determine these parameters empirically as in our previous works (Tao et al., 2019). The parameters *q*_{1} and *q*_{2} play the same role in optimizing ϑ_{a} and η_{a} for preventing the trivial solution of these optimal variables. Since the larger *q*_{1} (or *q*_{2} ) would lead to the same weights with greater probability, we therefore empirically set *q*_{1} = *q*_{2} = 2 in our experiments in terms of the suggestion provided in Hou et al. (2017). Besides this, we discreetly choose the values of λ, α, and β by employing the grid search strategy in a heuristic way. Concretely, these regularization parameters are tuned from {10^{−4},10^{−3}, …, 10^{3},10^{4}}. Finally, we search and fine-tune the number of nearest neighbors *k* in the set {3,5,10,15,17} for constructing the affinity graph in MACI (also in FSSL). For our algorithm, the maximum iteration number is set as τ = 100.

For those nonlinear learning methods MACI, FastDAM, and Multi-KT, we borrow the Gaussian kernel [i.e., *K*_{i,j} = exp (−σ||*x*_{i}−*x*_{j}||^{2})] as the default kernel function, where σ is determined by setting it to be the reciprocal of feature dimension 1/*d*. Following the same practice in Duan et al. (2012a), we predefine each source weight ${\mathrm{\gamma}}_{i}=\frac{\mathrm{exp}(-\mathrm{\delta}Dist({X}_{i}^{s},X))}{{\sum}_{i}\mathrm{exp}(-\mathrm{\delta}Dist({X}_{i}^{s},X))}$ ( *i* = 1, …, *S*) in FastDAM, where δ = 100.

#### Experiment I: Within-Dataset Emotion Recognition

It is worth noting that we may encounter difficulties with different subjects in EEG emotion recognition even if they belong to the same dataset because different subjects may have different EEG feature distributions due to personalized characteristics. Thereby, we may adopt the so-called leave-one-subject-out cross-validation strategy adopted also in Lan et al. (2018) to evaluate the performance of MACI on emotion recognition. Concretely, the left subject from the dataset of interest contributes to the target domain, and other subjects are constructed as the multi-source domains. We evaluate the multi-source adaptation performance of MACI compared with existing state-of-the-arts on SEED and DEAP, respectively.

There are totally 2,340 training data of 13 subjects with 60 data per class and 180 test data of each subject from three classes in DEAP. We extracted 2,775 samples consisting of 925 samples per class per class hour from each one of the 14 subjects in SEED, thus generating 38,850 training data from 14 subjects and 2,775 test samples from one target subject. Note that extant research (Chai et al., 2016, 2017; Zheng and Lu, 2016) have pointed out that it is almost impossible to train the DA methods by exploiting all training data from SEED due to the limitation of computational space. We thereby randomly sample 10% training data from SEED, i.e., 3,885 training data as the final multi-source domain data for all DA methods. We repeat each trial 10 times on SEED, and the final performance is the average of the results of 10 times.

#### Performance Comparison

We show in Table 2 the emotion recognition performance of MACI and several baselines on within-DEAP and within-SEED, respectively.

**Table 2.** Emotion recognition performance (mean % and SD %) of MACI and several baselines on within-datasets.

As reported in Lan et al. (2018), the theoretical performance (or chance level) of random guessing is about 33.33%, which could be approached by real chance level when the number of training samples increase to infinity (Lan et al., 2018). As shown in Table 1, the baseline FSSL contributes 40.17% mean recognition accuracy on DEAP, which is very near to the random value. When there are finite samples, we obtain the empirical chance level by repeating the trials of the samples in question equipped with randomized class labels (Lan et al., 2018). The finally obtained chance levels with bound of 95% confidence interval are also recorded in Table 2. We can see from Table 2 that the accuracy of FSSL significantly exceeds the upper bound of the real chance level at 5% significance level. However, the relatively lower performance of FSSL still indicates that emotion recognition with DA technique is imperative when there exists substantial divergence between the feature distributions of different subjects.

Almost all DA methods yield better recognition performance than FSSL for DEAP. Our MACI achieves the best performance (about 23.14% gains in performance over FSSL), closely followed by DSM. Note that though we acquired the relatively significant improvement measured by *t*-test with *p*-value > 0.05, the total recognition accuracy is still inferior. On SEED, FSSL achieves 53.78% average recognition accuracy over three sessions, which obviously surpasses the upper bound of the chance level. Several multi-source adaptation methods, i.e., Multi-KT, FastDAM, and DSM, undoubtedly obtain more performance gains than FSSL on SEED. The proposed method MACI still effectively boost the mean recognition accuracy up to 73.23% under *t*-test with *p*-value > 0.05, which still demonstrates the best performance on SEED. It is worthy to note that all methods including FSSL work more effectively on SEED than on DEAP. This interesting observation is partially consistent with that in Lan et al. (2018). A possible explanation may be that the so-called negative transfer prevented the effective application of DA techniques in DEAP since larger discrepancies among different subjects may exist in DEAP than in SEED (Mansour et al., 2009; Lan et al., 2018).

#### Number of Source Samples

Figure 1 presents the effect of varying the number of source samples. The source dataset size varies from 100 to 2,300 on DEAP and 100 to 3,800 on SEED, respectively. It can be seen from the curves in Figure 1 that all methods manifest the same trend of upgrade in the figure. This shows that larger source data is beneficial to improve the learning performance. It is worthy to note that the performance of MACI can be smoothly and steadily improved with the increase of the source samples, while other DA methods may only achieve satisfactory performance when the source samples are relatively large, i.e., larger than 500. In addition, A-SVM obtains the least performance on two datasets due to the so-called negative transfer issue in DA.

**Figure 1.** Classification accuracy with varying numbers of source samples on **(A)** DEAP and **(B)** SEED (session I).

On DEAP, MACI, and DSM maintain a better accuracy than other methods with less than 500 source domain samples. DSM outperforms Multi-KT and FastDAM when the number of target training samples is relatively large due to properly choosing the weights to assign to each subject. Our method MACI obtains even more gains over DSM when the number of source samples is increasing asymptotically. The reason may be that, except for the use of correlation information among subjects, MACI can effectively select the most related sources with the optimally weighted multi-source adaptation regularization. Besides that, partial experimental results also show that FastDAM could occasion the “negative transfer” issue with the MMD-based weights assigned to all sources, which would deteriorate its performance. On SEED, the accuracy flattens at above 3,000 source samples. When the number of source samples increases to 3,500, MACI, FastDAM, and Multi-KT asymptotically approach a similar performance. From this point onwards, MACI, FastDAM, and Multi-KT perform similarly if we have sufficient source data.

#### Multi-Kernel Learning

We further evaluate the effectiveness of our method with different kernel functions (called MKMACI for short) for each source domain. Given the empirical kernel mapping set ${\{{\mathrm{\varphi}}_{a}\}}_{a=1}^{\mathrm{\mho}}$, each mapping *X*_{a} into a different kernel space, we can integrate them orthogonally to the final space by concatenation, i.e., $\stackrel{~}{\mathrm{\varphi}}({x}_{i})={[{\mathrm{\varphi}}_{1}{({x}_{i})}^{T},{\mathrm{\varphi}}_{2}{({x}_{i})}^{T},\mathrm{\dots},{\mathrm{\varphi}}_{\mathrm{\mho}}{({x}_{i})}^{T}]}^{T}\in {R}^{\mathrm{\mho}{n}_{a}}$, for *x*_{i} ∈ *X*_{a}. The final kernel matrix in this new space is defined as *K*_{new} = [K̊_{1}; K̊_{2}; …; K̊_{℧}], where K̊_{i} is the kernel matrix in the *i*−th feature space. Therefore, besides the above-mentioned Gaussian kernel, we additionally employ another three types of kernels in MKMACI: Laplacian kernel ${K}_{ij}=\mathrm{exp}\left(-\sqrt{\mathrm{\sigma}}||{x}_{i}-{x}_{j}||\right)$, inverse square distance kernel *K*_{ij} = 1/(1 + σ||*x*_{i}−*x*_{j}||^{2}), and inverse distance kernel ${K}_{ij}=1/(1+\sqrt{\mathrm{\sigma}}||{x}_{i}-{x}_{j}||)$. It can be clearly seen in Figure 2 that MKMACI is obviously better than MACI in terms of mean accuracies in all cases, which justifies that the multi-kernel trick can improve the quality of DA emotion recognition on within-datasets.

### Experiment II: Cross-Dataset Emotion Recognition

Note that cross-dataset emotion recognition is more challenging in terms of the differences in acquisition and participant characteristics and behaviors. In the preceding experiments, we demonstrate the performance comparison of our method with other DA methods with the within-dataset (i.e., cross-subject) setting. We will, in this part, further evaluate the consistent effectiveness of MACI when performed on cross-dataset adaptation. In this scenario of experiment, we constructed multiple different schemes by sampling the training dataset and test dataset, respectively, with different EEG instruments and emotional stimuli. We therefore set up six trial settings, i.e., *DEAP* → *SEED I*, *DEAP* → *SEED II*, *DEAP* → *SEED III*, *SEED I* → *DEAP*, *SEED II* → *DEAP*, and *SEED III* → *DEAP*, to justify the effectiveness of MAC on cross-dataset emotion recognition. In the context, we denote A → B by adaptation from dataset A to dataset B. For simplicity of expression, we respectively, coined SEED I, SEED II, and SEED III as the dataset of session I, session II, and session III in the database SEED.

In universe DA, a commonly used hypothesis is that the feature space of both source and target domains is the same. Consequently, only 32 channels between SEED and DEAP are employed to formulate a 160 dimensional feature space for both training and test datasets. In the first three experimental settings, there are 180 × 14 = 2,520 source samples from DEAP and 2,775 target samples from three different sessions in SEED. We evaluate the recognition accuracy for each subject in each session and report the final experimental results based on the mean over 15 subjects from SEED. In the other experimental settings, a total of 2,775 × 15 = 41,625 source samples from SEED are regarded as training datasets, and 180 samples contributed from DEAP are test dataset. We then evaluate the recognition accuracy of individual subjects in DEAP, and the results are recorded with the average over 14 subjects. We randomly sample 10% of the source data (4,162 samples) as the actual training data due to the limitation of memory (Shi et al., 2013; Zheng et al., 2015; Chai et al., 2016, 2017; Zheng and Lu, 2016; Lan et al., 2018; Zhong et al., 2020). Under each setting, we conduct the trial repeatedly 10 times and record the average performance over these 10 times.

#### Performance Comparison

We record the mean experimental results on six cross-dataset settings in Table 3, from which we can observe that the performance of the baseline FSSL is inferior to the upper bound of chance level with 95% confidence interval, that is, the baseline performance is almost close to the random guess with 5% significance level. This indicates that there exist larger distribution divergences between two datasets as well as the variance among different subjects than that in within-dataset. The importance of DA would be indispensable in this scenario. This is justified by the observation in Table 3 that all DA methods outperform the baseline FSSL since DA could potentially reduce the technical discrepancies in cross-dataset applications. In most cases, MACI is found to be the best-performing method in the cross-dataset DA settings. In some scenario, Multi-KT and FastDAM occasionally obtain the best performance. A noticeable phenomenon can be observed in Table 3, such that the mean recognition accuracies of all methods are correspondingly worse than that in Table 2 obtained on within-dataset due to the larger distribution discrepancy between different datasets.

#### Multi-Source Adaptation

In practical DA applications, one may expect that the number of prior sources grow in time, which would incur the so-called scalability issue. In this problem, it is necessary to explore the reliability of each prior source for the specific task (Tao et al., 2019). To this end, we additionally conduct multi-source adaptation trials on several cross-dataset settings. The average results of MACI, DSM, Multi-KT, FastDAM, and A-SVM with the average prior model are reported in Figure 3.

**Figure 3.** Multi-source adaptation emotion recognition accuracy (SI, session I; SII, session II; SIII, session III).

It can be seen from the curves in Figure 3 that A-SVM is still worse than the other DA methods in most cases in that it is difficult for A-SVM to minimize the between-domain distribution distance when the distribution varies greatly between domains. The accuracies of A-SVM tend to be downgraded when the number of sources is increasing in some cases, suggesting that negative transfer may have happened in A-SVM. MACI obtains a relatively much better performance in most cases, which demonstrates that our algorithm can improve the emotion recognition performance on cross-dataset. All methods except A-SVM manifest the same trend of upgrade with the increase of sources, and the accuracy improvements are significant with respect to that of within-dataset settings. This shows that utilizing the limited sources is beneficial to improve the learning performance. In addition, MACI and DSM usually outperform other DA methods due to properly choosing the weights to assign to each source. Our method MACI obtains even more gains over DSM, which may be attributed to the utilization of correlation information among sources in that MACI can effectively select the most related source domains with the optimally weighted multi-source adaptation regularization.

#### Adaptation With Deep Features

In the past decade, deep learning attracts more and more attention due to its powerful representation ability and dramatic improvement over the traditional shallow methods. We therefore additionally compare our MACI method with the recently proposed deep transfer learning models DAN and ReverseGrad for cross-dataset emotion recognition using deeply extracted features in multi-source adaptation settings.

In our MACI, we can tackle the problem of deep DA with two steps: firstly, a higher-level feature extraction is learnt in an unsupervised fashion from all available domains using the popular deep architectures [e.g., VGG16 (Simonyan and Zisserman, 2014) or DAN]; secondly, our MACI is trained on the transformed data of all domains and then used to test the target domain. For fair comparison, however, we follow the experimental setup in Zhou et al. (2018) and Zhu et al. (2017). Specifically, we first fine-tune pretrained deep models (e.g., VGG16, DAN, and ReverseGrad) by using the labeled samples in the source domain and then use these fine-tuned CNN models to extract the features from EEG in both source and target domains. Finally, we perform emotion recognition using MACI on these deeply extracted features. In the context of our experiments, we denote our methods with different deep models as MACI + VGG16, MACI + DAN, and MACI + ReverseGrad, respectively. As for DAN and ReverseGrad, we use their released source codes and fine-tune the pre-trained deep models by using the suggested parameters in Long et al. (2015) and Ganin and Lempitsky (2015), respectively.

All experimental results are reported in Figure 4. As can be seen from this plot, the deep transfer learning methods are originally proposed to learn domain-invariant features, while our proposed method aims to improve the cross-domain generalization ability, namely, their methods focus on feature learning, while our work focuses on classification, so our proposed method can be used to further improve the recognition accuracies by co-learning the source classifiers with the features extracted by deep models, i.e., VGG16, DAN, and ReveseGrad. This indicates that the classification-level constraint can preserve all source discriminative structures for the guidance of target data classification, which demonstrates the effectiveness of MACI framework. From the plot bars of Figure 4, it can be observed that MACI + DAN consistently outperforms DAN, while MACI + ReverseGrad is consistently better than ReveseGrad, which demonstrates that our MACI method is complementary to the two deep transfer learning methods DAN and ReveseGrad by exploiting the correlation statistics to further enhance the generalization ability across domains.

**Figure 4.** Emotion recognition accuracies (%) of different methods using deeply extracted features (SI, session I; SII: session II; SIII, session III).

#### Parameter Impact on MACI

There are mainly three model parameters to be tuned in our method, i.e., λ, β, and α. Note that larger α would make the predicted label matrix better meet the expected needs, thus with better results being achieved. Consequently, we empirically set α = 10^{3} in the following experiments. We firstly explore to set the extreme values of different parameters for validating the importance of each component in our framework. Specifically, we denote MACI without the feature selection (i.e., β = 0) by MACI_NF and MACI with λ = η_{a} = 0 by MACI_NS, which ignores correlation information among multiple sources. These settings are evaluated on cross-dataset settings for multi-source adaptation tasks. From Table 4, we can observe that MACI can be significantly improved from MACI_NS by exploiting the correlation information among multiple sources. Besides this, the performance of MACI_NF is slightly weaker than MACI, that is, MACI would degrade when the feature selection function is omitted. A possible reason may be that the features of EEG represented by DE introduced some noise/outlier data. In this case, the feature selection in MACI possesses indispensable importance for robust DA learning. In sum, the utilization of correlation knowledge among sources and features could make MACI further boost its performance in cross-dataset emotion recognition applications. It is this argument that constitutes the basic principle of our MACI framework.

## Conclusion

In this work, we explore to cope with the cross-dataset emotion recognition where existing BCI methods cannot work well. To this end, we proposed an effective multi-source co-adaptation framework (MACI) for EEG-based emotion recognition mainly by leveraging correlation knowledge among sources and features in the objective function, which dampens unimportant evidence (within features and between sources) and amplifies useful knowledge. In MACI, multiple domain-invariant classification functions corresponding to different sources are co-learned by bridging both statistical and semantic distribution discrepancy between source and target domains, thus making MACI utilize the correlated knowledge among multiple sources by exploiting the developed correlation metric function. A large number of experimental results conducted on two publicly available EEG datasets show that MACI are much better than several representative baseline methods and provide the state-of-the-art performance on within/cross-dataset emotion recognition in most cases. This demonstrates the effectiveness of MACI in addressing feature distribution discrepancy between individual subjects as well as different datasets due to technical discrepancies.

To boost the efficiency of our method, however, a more efficient iterative algorithm would be developed or further elaborated in our future works. Besides this, the pseudo-labels strategy (i.e., iteratively updating target label matrix in the training stage) for bridging semantic distribution discrepancy between different domains would be unreliable or even misleading in training. This therefore arouses another challenge, i.e., how to effectively infer and incorporate target labels in unsupervised DA, which would be an urgent and valuable work in our future research.

## Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: http://epileptologie-bonn.de/cms/upload/workgroup/lehnertz/eegdata.html.

## Ethics Statement

Ethical review and approval was not required for the study on human participants in accordance with the local legislation and institutional requirements. Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements.

## Author Contributions

Both authors listed have made a substantial, direct and intellectual contribution to the work, and approved it for publication.

## Funding

This work was supported partly by High level talents introduction research project of Ningbo Polytechnic (under grant no. RC201902), partly by Zhejiang Provincial Natural Science Foundation of China (under grant no. LY19F020012), and partly by Foundation of Zhejiang Educational Committee (under grant no. Y201941140).

## Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

## Footnotes

**^**It is important to note that the feature mapping function ϕ_{a}(_{1≤a≤S}) with respect to each source domain can be completely different from each other.**^**For maximal consistency between*F*and^{T}F*I*_{c}, the parameter ζ should be set as a relatively large value. In our experiments, we therefore empirically set ζ = 10^{3}without loss of performance to some extent.**^**The code is available from http://vc.sce.ntu.edu.sg/transfer_learning_domain_adaptation_data/DAM-TNNLS2012.html.**^**We use the code available from http://homes.esat.kuleuven.be/~ttommasi/source_code_CVPR10.html.**^**The MATLAB code is available online http://www.robots.ox.ac.uk/~{}vgg/software/tabularasa/.**^**For each source domain, we train one SVM by using the corresponding labeled samples. Then, for each test instance*x*, the decision values from*p*SVM classifiers are converted into the probability values by using the sigmoid function [i.e., g(*t*) = 1/(1 + exp(−*t*)]. Finally, we average the*p*probability values as the final prediction of the test instance*x*.

## References

Bruzzone, L., and Marconcini, M. (2010). Domain adaptation problems: a DASVM classification technique and a circular validation strategy. *IEEE Trans. PAMI* 32, 770–787. doi: 10.1109/tpami.2009.57

Chai, X., Wang, Q., Zhao, Y., Li, Y., Liu, D., Liu, X., et al. (2017). A fast, efficient domain adaptation technique for cross-domain electroencephalography (EEG)-based emotion recognition. *Sensors* 17:1014. doi: 10.3390/s17051014

Chai, X., Wang, Q., Zhao, Y., Liu, X., Bai, O., and Li, Y. (2016). Unsupervised domain adaptation techniques based on auto-encoder for non-stationary EEG-based emotion recognition. *Comput. Biol. Med.* 79, 205–214. doi: 10.1016/j.compbiomed.2016.10.019

Chen, B., Lam, W., Tsang, I. W., and Wong, T.-L. (2013). Discovering low-rank shared concept space for adapting text mining models. *IEEE Transact. Patt. Anal. Mach. Intell.* 35, 1284–1297. doi: 10.1109/tpami.2012.243

Chu, W. S., Del, T. F., and Cohn, J. F. (2017). Selective transfer machine for personalized facial action unit detection. *IEEE Transact. Patt. Anal. Mach. Intell.* 39, 529–545.

Ding, Z., Sheng, L., Ming, S., and Fu, Y. (2018). “Graph adaptive knowledge transfer for unsupervised domain adaptation,” in *Proceedings of the 15th European Conference (ECCV2018), Munich, Germany, September 8-14, 2018*, (Cham: Springer).

Duan, L., Dong, X., and Shih-Fu, C. (2012a). “Exploiting web images for event recognition in consumer videos: A multiple source domain adaptation approach,” in *Proceedings of the CVPR, IEEE Computer Society Conference on Computer Vision and Pattern Recognition. IEEE Computer Society Conference on Computer Vision and Pattern Recognition CVPR*, (New York, NY: IEEE), 1338–1345.

Duan, L., Tsang, I. W., and Xu, D. (2012b). Domain transfer multiple kernel learning. *IEEE Transact. Patt. Anal. Mach. Intell.* 34, 465–479. doi: 10.1109/tpami.2011.114

Duan, L., Xu, D., and Tsang, I. W. (2012c). Domain adaptation from multiple sources: a domain-dependent regularization approach. *IEEE Transact. Neur. Netw. Learn. Syst.* 23, 504–518. doi: 10.1109/tnnls.2011.2178556

Ganin, Y., and Lempitsky, V. (2015). “Unsupervised domain adaptation by back-propagation,” in *Proceedings of the 32nd International Conference on Machine Learning*, Lille, 1180–1189.

Ganin, Y., Ustinova, E., Ajakan, H., Germain, P., Larochelle, H., Lavi-olette, F., et al. (2016). Domain-adversarial training of neural networks. *J. Mach. Learn. Res.* 17, 1–35. doi: 10.1109/tnnls.2020.3025954

Ghifary, M., Balduzzi, D., Kleijn, W. B., and Zhang, M. (2017). Scatter component analysis: a unified framework for domain adaptation and domain generalization. *IEEE Transact. Patt. Anal. Mach. Intell.* 99, 1–1.

Gretton, A., Fukumizu, K., Harchaoui, Z., and Sriperumbudur, B. K. (2009). “A fast, consistent kernel two-sample test,” in *Proceedings of the Conference on Neural Information Processing Systems 22*, (Vancouver, BC: MIT Press), 673–681.

Hou, C., Feiping, N., Hong, T., and Yi, D. (2017). Multi-view unsupervised feature selection with adaptive similarity and view weight. *IEEE Trans. Knowl. Data Eng.* 29, 1998–2011. doi: 10.1109/tkde.2017.2681670

Jayaram, V., Alamgir, M., Altun, Y., Scholkopf, B., and Grosse-Wentrup, M. (2016). Transfer learning in brain-computer interfaces. *IEEE Comput. Intell. Magaz.* 11, 20–31.

Jenke, R., Peer, A., and Buss, M. (2014). Feature extraction and selection for emotion recognition from EEG. *IEEE Transact. Aff. Comput.* 5, 327–339. doi: 10.1109/taffc.2014.2339834

Judy, H., Tzeng, E., Darrell, T., and Saenko, K. (2017). Simultaneous deep transfer across domains and tasks. *Dom. Adaptat. Comput. Vis. Appl.* 17, 173–187. doi: 10.1007/978-3-319-58347-1_9

Kim, M.-K., Kim, M., Oh, E., and Kim, S.-P. (2013). A review on the computational methods for emotional state estimation from the human EEG. *Comput. Mathematic. Methods Med.* 2013, 1–3. doi: 10.1155/2013/573734

Koelstra, S., Mühl, C., Soleymani, M., Lee, J. S., Yazdani, A., Ebrahimi, T., et al. (2012). DEAP: a database for emotion analysis using physiological signals. *IEEE Transact. Affect. Comput.* 3, 18–31. doi: 10.1109/t-affc.2011.15

Lan, Z., Sourina, O., Wang, L., Scherer, R., and Müller-Putz, G. R. (2018). Domain adaptation techniques for eeg-based emotion recognition: a comparative study on two public datasets. *IEEE Transact. Cogn. Dev. Syst.* 11, 85–94. doi: 10.1109/tcds.2018.2826840

Li, X., Song, D., Zhang, P., Zhang, Y., Hou, Y., and Hu, B. (2018a). Exploring EEG features in cross-subject emotion recognition. *Front. Neurosci.* 12:162.

Li, Y., Zheng, W., Cui, Z., Zhang, T., and Zong, Y. (2018b). “A novel neural network model based on cerebral hemispheric asymmetry for EEG emotion recognition,” in *Proceedings of the 27th International Joint Conference on Artificial Intelligence (IJCAI)*, Stockholm, 1561–1567.

Li, Y., Zheng, W., Cui, Z., Zong, Y., and Ge, S. (2018c). EEG emotion recognition based on graph regularized sparse linear regression. *Neur. Proces. Lett.* 8, 1–17. doi: 10.1109/taffc.2020.2994159

Li, Z., Liu, J., Tang, J., and Lu, H. (2015). Robust structured subspace learning for data representation. *IEEE Transact. Patt. Anal. Mach. Intell.* 37, 2085–2098. doi: 10.1109/tpami.2015.2400461

Long, M., Cao, Y., Wang, J., and Jordan, M. (2015). “Learning transferable features with deep adaptation networks,” in *Proceedings of the 32nd International Conference on International Conference on Machine Learning*, Lille, 97–105.

Long, M., Wang, J., Ding, G., and Pan, S. J. (2014). Adaptation regularization: a general framework for transfer learning. *IEEE Transact. Knowl. Data Eng.* 26, 1076–1089. doi: 10.1109/tkde.2013.111

Lotfi, E., and Akbarzadeh, M.-R. (2014). Practical emotional neural networks. *Neur. Netw.* 59, 61–72. doi: 10.1016/j.neunet.2014.06.012

Mansour, Y., Mohri, M., and Rostamizadeh, A. (2009). “Domain adaptation with multiple sources,” in *Proceedings of the Conference on Neural Information Processing Systems*, (Vancouver, BC: MIT Press), 1041–1048.

Mühl, C., Allison, B., Nijholt, A., and Chanel, G. (2014). A survey of affective brain computer interfaces: principles, state-of-the-art, and challenges. *Brain Comput. Interfac.* 1, 66–84. doi: 10.1080/2326263x.2014.912881

Musha, T., Terasaki, Y., Haque, H. A., and Ivamitsky, G. A. (1997). Feature extraction from EEGs associated with emotions. *Artif. Life Robot.* 1, 15–19. doi: 10.1007/bf02471106

Nie, F., Huang, H., Cai, X., and Ding, C. (2010a). “Efficient and robust feature selection via joint -norms minimization,” in *Proceedings of the International Conference on Neural Information Processing Systems*, (Vancouver, BC: Curran Associates Inc), 1813–1821.

Nie, F., Xu, D., Tsang, I. V.-H., and Zhang, C. (2010b). Flexible manifold embedding: a framework for semi-supervised and unsupervised dimension reduction. *IEEE Transact. Image Proces.* 19, 1921–1932. doi: 10.1109/tip.2010.2044958

Pan, S. J., Tsang, I. W., Kwok, J. T., and Yang, Q. (2011). Domain adaptation via transfer component analysis. *IEEE Transact. Neural Netw.* 22, 199–210. doi: 10.1109/tnn.2010.2091281

Pandey, P., and Seeja, K. (2019). “Emotional state recognition with EEG signals using subject independent approach,” in *Data Science and Big Data Analytics. Lecture Notes on Data Engineering and Communications Technologies*, Vol. 16, eds D. Mishra, X. S. Yang, and A. Unal (Singapore: Springer), 117–124. doi: 10.1007/978-981-10-7641-1_10

Rosenstein, M. T., Marx, Z., and Kaelbling, L. P. (2005). “To Transfer or not to transfer,” in *Proceedings of the Conference on Neural Information Processing Systems*, (Cambridge, MA: MIT Press).

Shi, L. C., Jiao, Y. Y., and Lu, B. L. (2013). “Differential entropy feature for EEG- based vigilance estimation,” in *Proceedings of the 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)*, Osaka, 6627–6630.

Simonyan, K., and Zisserman, A. (2014). “Very deep convolutional networks for large-scale image recognition,” in *Proceedings of the International Conference Learning Representations (ICLR)*, Banff, 1–14.

Song, T., Zheng, W., Song, P., and Cui, Z. (2018). EEG emotion recognition using dynamical graph convolutional neural networks. *IEEE Transact. Affect. Comput.* 18, 1–1.

Tao, J. W., Song, D., Wen, S., and Hu, W. (2017). Robust multi-source adaptation visual classification using supervised low-rank representation. *Patt. Recogn.* 61, 47–65. doi: 10.1016/j.patcog.2016.07.006

Tao, J., Chung, F. L., and Wang, S. (2012). On minimum distribution discrepancy support vector machine for domain adaptation. *Patt. Recogn.* 45, 3962–3984. doi: 10.1016/j.patcog.2012.04.014

Tao, J., Di, Z., Fangyu, L., and Bin, Z. (2019). Latent multi-feature co-regression for visual recognition by discriminatively leveraging multi-source models. *Patt. Recogn.* 87, 296–316. doi: 10.1016/j.patcog.2018.10.023

Tao, J., Wen, S., and Hu, W. (2015). L1-norm locally linear representation regularization multi-source adaptation learning. *Neural Netw.* 69, 80–98. doi: 10.1016/j.neunet.2015.01.009

Tao, J., Wen, S., and Hu, W. (2016). Multi-source adaptation learning with global and local regularization by exploiting joint kernel sparse representation. *Knowl. Based Syst.* 98, 76–94. doi: 10.1016/j.knosys.2016.01.021

Tommasi, T., Orabona, F., and Caputo, B. (2014). Learning categories from few examples with multi model knowledge transfer. *IEEE Trans. Pattern Anal. Mach. Intell.* 36, 928–941. doi: 10.1109/tpami.2013.197

Tzeng, E., Hoffman, J., Saenko, K., and Darrell, T. (2017). “Adversarial discriminative domain adaptation,” in *Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)*, Honolulu, HI, 2962–2971.

Wang, C., and Mahadevan, S. (2011). “Heterogeneous domain adaptation using manifold alignment,” in *Proceedings of the International Joint Conference on Artificial Intelligence*, Barcelona, 1541–1546.

Yan, S., Xu, D., Zhang, B., Yang, Q., and Lin, S. (2006). Graph embedding and extensions: a general framework for dimensionality reduction. *IEEE Transact. Patt. Anal. Mach. Intell.* 29:40. doi: 10.1109/tpami.2007.250598

Yang, J., Yan, R., and Hauptmann, A. G. (2007). “Cross-domain video concept detection using adaptive svms,” in *Proceedings of the ACM International Conference on Multimedia*, (New York, NY: ACM), 188–197.

Yang, Y., Ma, Z., Hauptmann, A. G., and Sebe, N. (2013). Feature selection for multimedia analysis by sharing information among multiple tasks. *IEEE Transact. Multim.* 15, 661–669. doi: 10.1109/tmm.2012.2237023

Zhang, K., Gong, M., and Schölkopf, B. (2015). “Multi-source domain adaptation: a causal view,” in *Proceedings of the 29th AAAI Conference on Artificial Intelligence*, Austin, TX, 3150–3157.

Zhang, Y., Chung, F. L., and Wang, S. (2019a). Takagi-sugeno-kang fuzzy systems with dynamic rule weights. *J. Intell. Fuzzy Syst.* 37, 8535–8550. doi: 10.3233/jifs-182561

Zhang, Y., Chung, F., and Wang, S. (2020a). Clustering by transmission learning from data density to label manifold with statistical diffusion. *Knowl Based Syst.* 193:105330. doi: 10.1016/j.knosys.2019.105330

Zhang, Y., Dong, J., Zhu, J., and Wu, C. (2019b). Common and special knowledge-driven TSK fuzzy system and its modeling and application for epileptic EEG signals recognition. *IEEE Access* 7, 127600–127614. doi: 10.1109/access.2019.2937657

Zhang, Y., Li, J., Zhou, X., Zhang, M., Ren, J., and Yang, J. (2019c). A view-reduction based multi-view TSK fuzzy system and its application for textile color classification. *J. Amb. Intell. Human. Comput.* 19, 1–11.

Zhang, Y., Tian, F., Wu, H., Xingyun, G., Danmin, Q., Jiancheng, D., et al. (2017). Brain MRI tissue classification based fuzzy clustering with competitive learning. *J. Med. Imag. Health Inform.* 7, 1654–1659. doi: 10.1166/jmihi.2017.2181

Zhang, Y., Wang, L., Wu, H., Geng, X., Yao, D., and Dong, J. (2016). A clustering method based on fast exemplar finding and its application on brain magnetic resonance images segmentation. *J. Med. Imag. Health Inform.* 6, 1337–1344. doi: 10.1166/jmihi.2016.1923

Zhang, Y., Wang, S., Xia, K., Jiang, Y., and Qian, P. (2020b). Alzheimer’s disease multiclass diagnosis via multimodal neuroimaging embedding feature selection and fusion. *Inform. Fus.* 66, 170–183. doi: 10.1016/j.inffus.2020.09.002

Zheng, W. (2017). Multichannel EEG-based emotion recognition via group sparse canonical correlation analysis. *IEEE Transact. Cogn. Dev. Syst.* 9, 281–290. doi: 10.1109/tcds.2016.2587290

Zheng, W. L., and Lu, B. L. (2015). Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. *IEEE Transact. Autonom. Mental Dev.* 7, 162–175. doi: 10.1109/tamd.2015.2431497

Zheng, W. L., and Lu, B. L. (2016). “Personalizing EEG-based affective models with transfer learning,” in *Proceedings of the 25th International Joint Conference on Artificial Intelligence*, Palo Alto, CA, 2732–2738.

Zheng, W. L., Zhang, Y. Q., Zhu, J. Y., and Lu, B. L. (2015). “Transfer components between subjects for EEG-based emotion recognition,” in *Proceedings of the 2015 International Conference on Affective Computing and Intelligent Interaction (ACII)*, Xi’an, 917–922.

Zhong, P., Wang, D., and Miao, C. (2020). EEG-based emotion recognition using regularized graph neural networks. *IEEE Transact. Affect. Comput.* 99, 1–1.

Zhou, X., Jin, K., Shang, Y., and Guo, G. (2018). Visually interpretable representation learning for depression recognition from facial Im-ages. *IEEE Transact. Affect. Comput.* 11, 542–552. doi: 10.1109/TAFFC.2018.2828819

Keywords: electroencephalogram, emotion recognition, multi-source adaptation, feature selection, maximum mean discrepancy

Citation: Tao J and Dan Y (2021) Multi-Source Co-adaptation for EEG-Based Emotion Recognition by Mining Correlation Information. *Front. Neurosci.* 15:677106. doi: 10.3389/fnins.2021.677106

Received: 07 March 2021; Accepted: 22 March 2021;

Published: 13 May 2021.

Edited by:

Yuanpeng Zhang, Nantong University, ChinaReviewed by:

Hongjiang Wei, Shanghai Jiao Tong University, ChinaCai Zhenhua, Wuhan University of Technology, China

Copyright © 2021 Tao and Dan. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jianwen Tao, 2019025@nbpt.edu.cn

^{†}These authors have contributed equally to this work