- 1Country Computer Integrated Manufacturing System Research Center, College of Electronics and Information Engineering, Tongji University, Shanghai, China
- 2Tsinghua Shenzhen International Graduate School, Tsinghua University, Shenzhen, China
The entity relation extraction in the form of triples from unstructured text is a key step for self-learning knowledge graph construction. Two main methods have been proposed to extract relation triples, namely, the pipeline method and the joint learning approach. However, these models do not deal with the overlapping relation problem well. To overcome this challenge, we present a relation-oriented model with global context information for joint entity relation extraction, namely, ROMGCJE, which is an encoder–decoder model. The encoder layer aims to build long-term dependencies among words and capture rich global context representation. Besides, the relation-aware attention mechanism is applied to make use of the relation information to guide the entity detection. The decoder part consists of a multi-relation classifier for the relation classification task, and an improved long short-term memory for the entity recognition task. Finally, the minimum risk training mechanism is introduced to jointly train the model to generate final relation triples. Comprehensive experiments conducted on two public datasets, NYT and WebNLG, show that our model can effectively extract overlapping relation triples and outperforms the current state-of-the-art methods.
Introduction
Relation extraction (RE) is a significant task for constructing self-learning knowledge graphs (KGs), which are graph-structured facts usually in the form of triple. The relationship between an entity pair (ei, ej) can be formalized as a relational triple (ei, rij, ej), where ei and ej represent the head and tail entity, respectively, and rij denotes a specific type of relationship connecting ei to ej. The objective of RE is to extract relation triples from unstructured text without human intervention based on predefined entity and relation categories. The well-structured nature of relational triples is well-suited to develop KG, which is widely used in intelligent robot, intelligent recommendation, intelligent furniture, and so on. In particular, products with question-and-answer functions such as chatbots (e.g., Siri, Microsoft Xiaoice, Tmall Genie, Xiaomi speakers, etc.) require the support of large-scale KG. However, the RE model with bad performance will lead to incomplete KG (Wu and Luo, 2020; Liu et al., 2021). Therefore, it is essential to construct effective RE models to extract all triples from texts.
Conventional RE task is based on pipeline methods in which the named entity recognition (NER) part is first applied to recognize entities in a sequence (Liu et al., 2010; Vazquez et al., 2011; Skeppstedt et al., 2014), and then a relation classification (RC) part is used to assign the predefined relation types to these candidate entity pairs (Santos et al., 2015; Zhang et al., 2018; Ren et al., 2022). After these two sequential steps, triples are finally extracted. Although such a structure makes the task simple to conduct, it ignores the hidden interdependency and error propagation between these two subtasks, which leads to low accuracy in extracting relation triples. To overcome these problems, joint extraction methods, simultaneously detecting entities together with their relations from unstructured texts, are proposed.
As shown in Table 1, sentences are generally classified into three types according to the overlapping degree of triples (Zeng et al., 2018), namely, (1) normal: there is no entity belonging to two or more different triples simultaneously in a sentence; (2) entity pair overlap (EPO): there is entity pair (ei, ej) sharing two or more different relations, i.e., , in a sentence. (3) single entity overlap (SEO): there is an entity ei that has two or more relations with different entities in a sentence, i.e., (ei, rij, ej), (ei, rik, ek), ⋯, (ei, riq, eq). Due to the existence of complex overlapping relations (e.g., SEO or EPO), the joint extraction methods still face great challenges in extracting relation triples.
Recently, extensive research works have been conducted in joint extraction. These works are divided into two directions based on the extraction order of the triple elements: entity first and relation first. The entity-first method can be formed as (ei, ej) → rij, which first identifies all entities applying NER techniques, and then assigns relation types to these candidate entity pairs. Yu et al. (2019) proposed a new model ETL–Span, where the head entities were first identified, and then the matching tail entities and relations were recognized by using some joint decoding approaches. Li et al. (2019) developed a novel paradigm by casting the RE task as a multi-turn question answering problem, i.e., the extraction of entities and relations was changed to the task of recognizing answer spans from the context. In this process, the head entity was first extracted, and then the tail entity and the relation were detected.
The relation-first method is formed as rij, → (ei, ej), where the relation information can be used as the prior knowledge to guide the extraction of semantically related entities. In the CopyRE model (Zeng et al., 2018), the relation was first generated by the decoder with the copy mechanism. Then the copy mechanism was adopted to extract the head entity and tail entity from the source text. However, CopyRE cannot distinguish head and tail entities or predict multi-token entities (e.g., Steven Jobs). To solve these problems, Takanobu et al. (2019) proposed the hierarchical reinforcement learning (HRL) framework, where a high-level reinforcement learning (RL) was used to detect the relations, and a low-level RL was applied to identify the participating entities related to the detected relations.
Although the above approaches have achieved reasonable performance in extracting relation triples, they all suffer from the same problem, namely, exposure bias. Besides, these models neglect the semantic connections between words in a text. Therefore, the problem of extracting overlapping relations is not overcome. To this end, this study proposes a novel relation-oriented model with global context information for joint extraction (ROMGCJE), which can effectively extract the overlapping relation triples from the unstructured texts. As illustrated in Figure 1, this model is constructed in the encoder–decoder structure. The encoder part consists of a primary task-shared representation layer (PTSRL), a contextual word representation layer (CWRL), and a relation-based attention module with a gating mechanism (RBAGM). First, a tagging scheme (Liu et al., 2019) is applied to convert the joint extraction task to a sequence labeling problem, mapping a sentence into a tag sequence . Then, the PTSRL, composed of BERT (Devlin et al., 2018) and bidirectional long short-term memory neural network (BiLSTM) (Schuster and Paliwal, 1997), is used to obtain word embedding. Next, the CWRL, made of a multi-head attention module (MHAttention) and multi-layer dilated convolution (MDiconv) module with residual blocks, is proposed to encode words to capture rich global contextual features, build long-range dependency among words, and more fully extract the semantic connections between words. Furthermore, the RBAGM is designed to utilize the relation information to guide the detection of entities. The encoder layer makes the ROMGCJE model jointly express both entities and relations with shared parameters in a single model. The decoder part contains a multiple relation classifier (MRC) (Read et al., 2011) for extracting relations and an improved LSTM (MGLSTM) for detecting entities. Furthermore, an attention module is used to match the corresponding entities in line with the recognized relations. Eventually, the minimum risk training (MRT) mechanism (Sun et al., 2018) is introduced to train the model to make the training process more stable. This approach does not desire any supplementary manual features or natural language processing (NLP) tools compared with feature-based methods. In addition, this new model not only prevents superfluous information from being generated but also tackle overlapping relations effectively.

Figure 1. The structure of ROMGCJE. The tokens of sentences are first represented as distributed word representations, which are then fed to the encoder layer containing PTSRL, CWRL, and RBAGM to capture rich contextual information and build long range dependency. Next, MGLSTM and MRC in the decoder layer are applied to perform decoding for NER and RC tasks, respectively. Finally, the MRT mechanism is applied to train ROMGCJE to extract the final relation triples.
The contributions of this study are summarized as follows: (1) A novel relation-oriented model with global context information is proposed for joint extraction. The relation-aware attention mechanism is applied to make use of the relation information to guide the entity detection, which can reduce the extraction of redundant entities. (2) This model takes into consideration the rich global contextual information, builds long-range dependency among words, and fully extracts the semantics of the passage in extracting relation triples. (3) Extensive experiments conducted on public datasets (NYT and WebNLG) demonstrate that the proposed model can achieve state-of-the-art performance in extracting overlapping relation triples.
Related Work
Relation extraction has become progressively critical in KG construction, smart robots, search engine, intelligent question–answering systems, etc. The pipeline method and joint method are the mainstream methods for extracting entities and relations from unstructured texts.
In early works, the pipeline method is intensively investigated, which divides the task into two serial independent subtasks, NER and RC; NER, recognizing the named entities, generally, is defined as a sequence tagging task. There are two research directions, i.e., statistical method (Gong et al., 2019; Alam and Islam, 2020) and neural network method (NNM) (Wei et al., 2019; Zhang et al., 2019). For the statistical machine learning-based method, the conditional random fields (CRF) mechanism is widely used, and the feature engineering and corpora play critical roles in improving extraction accuracy (Dalvi et al., 2011). Neural network method typically uses neural networks, e.g., convolutional neural networks (CNNs) or recurrent neural networks (RNNs), to learn sentence features without tedious feature engineering. There are some classic models, such as BiLSTM + CRF (Huang et al., 2015), deep neural networks +R (Tomori et al., 2016), and so on. In addition, it can achieve outstanding results by combining with the pre-trained language models such as BERT (Dai et al., 2019a). Relation classification, determining the relationship type between two entities, generally, is taken as a multi-label classification task. Zhou et al. (2016) introduced the attention mechanism to determine the contribution of each word to the final results. Zhang et al. (2017) applied the position-aware attention mechanism to refine the information of position embedding. However, the above approaches require preprocessing step NER, which leads to that the errors generated in the NER stage will be propagated to the RC stage (Zeng et al., 2018). Besides, NER task and RC task are realized by different models without joint training, which leads to that the inherent correlation between NER task and RC task is neglected and extensive unrelated entity pairs are generated in the pairing phase (Zhao et al., 2021).
To overcome the limitations of pipeline methods, a number of joint extraction methods have been proposed (Bekoulis et al., 2018). Zheng et al. (2017) presented a novel tagging scheme to convert the extracting task into a sequence labeling problem, which could make use of the inherent correlation between the RC task and the NER task. Nevertheless, this approach can only extract triples from normal sentences where there is no entity belonging to two or more different triples simultaneously. To extract overlapping relation triples, Zeng et al. (2018) put forward an end-to-end model with the copy mechanism, but the NER part heavily depended on word segmentation tools. Bekoulis et al. (2018) proposed a joint training model, where the NER task was realized by the CRF layer and the RC task was taken as a multi-head selection problem. Fu and Ma (2019) put forward a new joint model based on the LSTM and graph convolutional networks (GCNs). However, the above approaches only deal with the SEO problems and fail to solve the EPO problems, since they cannot assign different relation tags to one token.
To overcome the EPO problems, great efforts have been made. Li et al. (2021) proposed a translating decoding schema for joint extraction of entities and relations (TDEER), but this model did not deal with the error accumulation problem. To solve this problem, Wang et al. (2020a) transformed the joint extraction task into a token pair linking (TPLinker) problem, which did not contain any interdependent stages. Although the TPLinker could alleviate the error accumulation problem, processing all token pairs at encoder layers led to high computation complexity in encoding long paragraphs. Chen et al. (2019) proposed a novel architecture, where a BiLSTM classifier was first applied for identifying all possible relations maintained in the text, and then multi-head attention was performed to generate all possible entity pairs sequentially. Eberts and Ulges (2019) developed an attention model for span-based joint entity and relation extraction. The model achieved excellent performance but still suffered from the redundancy problem. Dai et al. (2019b) proposed a tagging scheme to produce m tag sequences for a sentence with m words, and applied a position-attention mechanism to generate different sentence representations for each query position. The joint extraction task was decomposed into the following two subtasks as in the study by Wang et al. (2020b): Head entity extraction and tail entity extraction, which solved the problem of SEO in the triples. Then the tail entity extraction was divided into three parallel sub-processes to solve the EPO problem of triples. Nevertheless, because of the labeling-once process, the above models always ignore the inner dependency among head entities, tail entities, and relations. Luo et al. (2020) proposed a bidirectional tree tagging scheme to label overlapping triples in text, which promoted the extraction of overlapping relation triples to some extent. Unfortunately, these models failed to extract overlapping relation triples in high accuracy, since they do not take into consideration the underlying contextual information or the semantic relation between words during extracting overlapping relation triples. Accordingly, the overlapping relation problem is not handled. Based on the above analysis, we propose a new relation-oriented model with global context information for the joint extraction of overlapping relations and entities. This model can deal with the overlapping relations effectively, since it can take full advantage of rich global contextual information, build long-range dependency among words, and more fully extract the semantics of the passage in extracting relation triples.
Model Description
In this section, the task description is first introduced. Then we describe the tagging method and explain how to change the extraction task to a tagging problem. Finally, the new model ROMGCJE is introduced in detail, which can extract overlapping relation triples effectively.
Task Description
The objective of this new model is to extract all entities together with their relations in the form of triples from unstructured text. The relational triple is formed as (ei, rij, ej), where ei ∈ E, ej ∈ E, rij ∈ R. Denote E and R as a set of predefined entities and relations, respectively. Especially, some entities or relations may exist in multiple triples. The entity extraction is taken as a sequence tagging task, where different tags are assigned to different words in the input sequence. The RC task is taken as a multi-label classification task. During the extraction process, entity information and relation information can interact.
Tagging Method
To overcome the problem of overlapping relations, the Beginning, Inside, End, Outside, and Single (BIEOS) scheme is applied to label entities and relations in sentences. The word's tag contains information about the word's position in the entity, the relation type, and the relation role. The word position information refers to BIEOS. All relation types come from the predefined set of relations, and the numbers “1” and “2” are used to represent the relation role information. “1” represents that the word belongs to the first entity in the triple, while “2” denotes that the word belongs to the second entity in the triple.
Figure 2 explains how the sentences are tagged by using the BIEOS scheme (Liu et al., 2019). In one sentence, multiple tag sequences are given, and each tag sequence only contains a triple. Therefore, even though there exist overlapping relations in a sentence, entities can be assigned to the right labels. The input sentence contains three triples: {Joe Biden, Country-President, America}, {Joe Biden, Nationality, America}, and {Capitol Hill, Contains, Washington}, where “Country-President,” “Nationality,” and “Contains” are the predefined relation types. The words “Joe,” “Biden,” “America,” “Capitol,” “Hill,” and “Washington” are related to the predefined relation types. For instance, the word “Joe” is the first word of the entity “Joe Biden” and is related to the relation “Country-President,” so its tag is “B-CP-1.” The other entity, “America,” which corresponds to “Joe Biden,” is labeled as “S-CP-1.” Besides, other words irrelevant to the final result are labeled as “‘O.” We combine entities with the same relation type into a triple to get the final result. Specifically, “Joe Biden” and “America” can be combined into a triple whose relation type is “Country-President.” Because the relation role of “Joe Biden” is “2” and “America” is “1,' the final result is {America, Country-President, Joe Biden}. Besides, if a sentence contains two or more triples with the same relation type, we combine every two entities into a triple based on the nearest principle. For example, if the relation type “Country-President” in Figure 2 is replaced by “Contains,” then there will be four entities with the same relation type in the given sentence. “America” is closest to the entity “Joe Biden,” and the “Capitol Hill” is closest to “Washington,” so the final results will be {America, Contains, Joe Biden} and {Capitol Hill, Contains, Washington}.
Evidently, the entities can be used multiple times in the different relation triples. Inspired by Miwa and Bansal (2016), the random start and fine-turned operations are applied to these tag labels during training. Notice that the ground-truth labels are only used during training, whereas the predicted labels are used at inference time.
Encoder Layer
The encoder layer, composed of PTSRL, CWRL, and RBAGM, is designed to better capture global contextual information, build long-range dependency among words, and fully extract the semantics of text. Note that we add a [CLS] token in front of the sequence and a [SEP] token at the end of the sequence.
PTSRL
As illustrated in Figure 1, the PTSRL consists of a BERT pretraining model and a BiLSTM. Given a sentence that consists of n words , where wt represents the tth word, we map each token in the sentence to a real-valued embedding to express its semantic and syntactic meaning through the BERT word embedding layer, and get by Equation (1),
where represents the d-dimensional word vector embedded to the tth word in the sentence.
As out of vocabulary word is common for entity, we also augment word representation with character-level information. A BiLSTM network, as illustrated in Figure 1, is applied to obtain the character-level representations by Equation (2), which effectively captures the morphological information of the word.
The primary task-shared representations containing the word-level semantic information are the concatenation of word-level and character-level representation.
Contextual Word Representation Layer
In the multi-label classification task, nT relation types represent nT classes. Each relation type in one sentence has semantic units that constitute the entire text's semantic meaning. The primary task-shared representations are not enough for encoding the nT tag sequences in a sentence. Therefore, we design the CWRL, which is composed of the MDiconv module, MHAttention module, and Max-Pooling module, to capture semantic units, build long sequence information dependence, and extract rich global contextual information.
MDiconv Module
Plenty of research works have proved that the dilated convolution (Diconv) performs well in expanding the receiving field without losing position and semantic information. Inspired by Salimans and Kingma (2016), we design the Mdiconv module. As we all know, the deeper network usually can extract more abstract features and richer semantic information. However, simply increasing the network depth will lead to gradient explosion and gradient disappearance. To deal with this problem, the skip connection is introduced to the Mdiconv module. As shown in Figure 3, there are three residual blocks. Each residual block is composed of the dilated causal convolution, weight normalization, rectified linear unit (ReLU), and spatial dropout for regularization. A 1 × 1 convolution is introduced into each residual block, since there is some difference in dimension between the input and output. After the elementwise addition ⊕ operator, the features are fed to the next residual block, and finally we obtain the final sentence representations . Because of the introduction of the skip connection, MDiconv exhibits longer memory, which makes it suitable for encoding long sentences containing entities far from each other.
MHAttention Module
The multi-head self-attention is applied to capture arbitrary interactions between tokens. It has several merits as follows: (1) Building long-range dependencies by explicitly attending to all positions (2) disambiguating homonyms, and expressing semantic and syntactic patterns greatly. Specifically, the hidden features of the MHAttention module are calculated as follows.
where different linear layers are used in Equations (3)–(5) to map the input O into different subspaces by learnable parameters , and yielding query Q, keys K, and values V. Then z parallel heads are employed to develop interactions in the different parts of channels to generate text representation H. The representation headj of the jth head is calculated by a scaled dot-product attention operation in Equation (7).
Then, a residual connection along with layer normalization is applied on O and H to generate the output hidden features . Finally, a max-pooling layer is appended on all hidden states of the MHAttention module to capture a sentence-level feature .
where maxp represents the max-pooling operation.
The sentence feature representations M is obtained by Equation (9), which encodes the semantic information of its context.
where Norm represents a normalization operation.
Relation-Based Attention Module With a Gating Mechanism
The different words in a sentence with the different relations play the different roles. Therefore, relation-based attention (RBA) (Yuan et al., 2020) is applied to help calculate the sentence representations. The attention weight is proportional to the influence of the word at the current decoding time. When considering the relation type rk, the specific context sentence representation ck is calculated by the weighted sum of the sentence words, defined by Equations (10)–(12).
where weight matric Wh, Ws, and Wk, and bias vector battn are the parameters; mi denotes the hidden state vector of the ith word in the encoder layer; M represents the global representation of the sentence; rk indicates the trainable embedding of the kth relation; αt, i represents the RBA score, which can weigh the importance of each word to the relational expression greatly.
Only under the circumstance that the relation is positive to the sentence do the relation-oriented representations make sense to the following entity extraction. To adaptively control the relation information provided by the previous attention layer, the gated operation mechanism (GOM), which is defined as follows, is applied:
where W1, W2, W3, b1, b2, and b3 represent learnable parameters, and ⊙ is the dot product. σ indicates the elementwise sigmoid activation function, which returns values from 0 to 1. Therefore, the final results are taken as the percentage of information to maintain. Equation (14) aims to weigh whether the inherent sentence representation M or the relation-based representation ck is more effective for entity extraction; uk represents the reserved relational feature. The final representation of the ith word is obtained by concatenating mi and uk as follows:
The sentence is thus represented as .
Decoder Layer
The decoder part contains the MRC for extracting relations and the MGLSTM for detecting entities. An attention module is applied to match the corresponding entities in line with the identified relations. Finally, the MRT mechanism is introduced to train the model to generate triples. In the following part, the composition of the decoder part will be illustrated in detail.
Multiple Relation Classifier
The relation prediction task is a multi-label classification task, aiming to recognize all relation types in text. Commonly, one sentence includes multiple relation triples, which have connections to each other. The MRC is introduced for relations prediction, effectively preventing the phenomenon from happening that multiple classifiers predict the same relation.
To train better relation classifiers to improve the classification accuracy, the output vector Mk of the encoder layer and the global embedding G in MGLSTM are fused to constructure the relation layer .
Then a convolution operator Conv and a max-pooling operator are applied on the relation layer δ to generate the text embedding ρ:
where φ ∈ ℝm×(n−l+1), m represents the number of different filters, n denotes the length of the text, and l indicates the convolutional filter size.
The binary classifier for the jth relation type is defined as follows:
where Rj represents the relation embedding for jth relation type, and denotes the probability distribution that whether the text contains the jth relation type or not; and are learnable weight parameters. If the text contains the jth relation type, the relation embedding Rj will be fed into the joint extraction module (AJE) to aid entity pair recognition.
The training objective is to minimize the loss function Lrel
where Wr represents the weight matrix, br denotes the bias vector, indicates the predicted relation, and Rj is the real relation.
MGLSTM
The entity extraction is taken as a sequence labeling task, which focuses on correctly detecting and identifying all entities with relations in one sentence S. During this process, we do not determine the relation type between two entities. To realize this objective, the MGLSTM is applied, as shown in Figure 4, which is an excellent variant of LSTM. It inherits the most characteristics of RNN models and solves the problem of gradient disappearance in the gradient backpropagation process. The detailed operations are defined as follows:
where denotes the input, g(yt−1) indicates the global embedding, ct−1 represents attention context at time step t − 1, and σ denotes the logistic sigmoid function. In detail, the forget gate ftcontrols how much the previous memory cell is forgotten, the input gate it controls how much information is input to each unit, and the output gate otcontrols the exposure of the internal memory state. denotes the output which can predict all entities in the text.
Commonly, the predicted labels in the previous steps have some influence on the label prediction in the following steps. This is a classical exposure bias phenomenon, which might occur in all time steps. To overcome this challenge, all information at the time step t − 1 should be considered. Therefore, the global embedding g(yt−1) is introduced, which represents the information of all candidate labels that exist in yt−1 at the (t−1)th time step; yt−1 represents the predicted probability distribution at time step t − 1. The global embedding g(yt−1) is calculated as follows:
where e represents the embedding of the label which has the biggest probability under the distribution yt−1. denotes the weighted average embedding at time t; represents the ith element of yt−1, and ei denotes the embedding vector of the ith label; γ denotes the transform gate used for regulating the proportion of the weighted average embedding. Global embedding g(yt−1) is the optimized combination of the original embedding and the weighted average embedding with transformed gate γ. The global embedding improves the performance of the model significantly, since the source information is optimized when predicting the label at time step t.
In addition, the label sequence of each sample is sorted based on the frequency of labels in the training set. To predict the tag , a Softmax layer is applied to get the final probability distribution Pent over the label space ϑ, which is computed as follows.
where Wo, Wh, Wc, and Wsf denote the weight matrices, and bsf represents the bias vector; It is the mask vector which aims to prevent the decoder from generating repeated labels; φ represents a non-linear activation function.
In the training process, the cross-entropy loss is taken as the loss function, and the beam search algorithm (Wiseman and Rush, 2016) is applied to figure out the top-ranked prediction path. Given an input sentence S and its ground-truth tag sequence Tag, the training objective is to minimize loss function Lent.
where Wall denotes weight matrices, ball represents bias vector, indicates the predicted tag, and Tag stands for the real tag.
Joint Extraction of Entities and Relations
After detecting all relations and all entities, the extraction of triples is realized by applying the AJE. During this process, two entities are selected as the target entity pair, and the different entity pairs are assigned the different target relations. The weighted value generated by the attention mechanism represents the matching degree between the token and the target relation type. To get the attention weight of the ith predicted relation, the annotation sequence of the ith relation at time step t is calculated as follows:
where oci indicates the trainable embedding of the ith relation. Gt−1 denotes the global embedding generated by the entity prediction at time step t − 1; mi represents the embedding generated by the encoder. Wr, Wd, and Wp represent learnable parameters.
The attention distribution αt on the annotation sequence βi of text is computed as follows:
where , n denotes the text length. denotes the output at time step t.
Finally, the label sequence corresponding to the ith relation is calculated by follows:
where denotes the input vector at time step t and represents the context vector. The sequence vector generated for the ith relation is .
Training
The MRT Framework
To make the training more stable, the MRT framework is introduced. We pretrain the model with local loss in the first step, and we optimize the local loss Llocal and the global loss Lmrt simultaneously in the second step. The local loss Llocal, defined in Equation (41), is the linear combination of the MRC loss Lrel and the MGLSTM lossLent:
where λ is a hyperparameter to balance the RC task and the NER task.
The global loss Lmrt provides a tighter connection between the entity extraction task and the relation classification task. To illustrate the algorithm, we first aggregate some notations. Let y ≜ (Tag, R) contain the ground truth entity tag sequence and relations, contain the predicted entity tag sequence and relations, and ∅(S) be the set of all possible outputs of the input sentence S [y, ŷ ∈ ∅(S)]. We define the joint probability by
where W = Wst ∪ WR represents the set of the model parameters. The MRT loss Lmrt is defined by
where Q(ŷ|S; W, u, αmr) is a re-normalization of P(ŷ|S; W) on the subset ∅′(S). The sharpness of Q distribution (Och, 2017) is regulated by hyperparameter αmr, and u measures the significance of the entity model and the relation model in Q; ntri represents the number of triples.
Training Algorithm
The complete training step for ROMGCJE is summarized in Algorithm 1. The embedding dimensions, learning rate, dilation rates, the number of heads, and so on., will be initialized before training. During the training process, the mini-batch samples are first fed into the encoder layer, which contains PTSRL, CWRL, and RBAGM. Then we obtain the hidden states of the encoder, which contains rich contextual information and semantic information. Next, the MRC and MGLSTM in the decoder layer are applied to complete the NER and RC tasks. After the detection of all relations and all entities, the label sequences of triples are realized by the application of the AJE. Finally, the whole model is trained by the MRT framework, which updates all parameters based on the mini-batch gradient descent.
Experiments
To evaluate the effectiveness of the new model, extensive experiments are conducted on the two different public datasets NYT and WebNLG.
Dataset
The NYT dataset was obtained based on the distant supervision method without artificial labeling (Riedel et al., 2010). WebNLG dataset was introduced by Gardent et al. (2017) for the natural language generation task and later was applied to the triple extraction task. The summary statistics of the two datasets are shown in Table 2. The test set is divided into different groups according to the triple overlapping degree and the number of triples in one sentence.
Hyperparameters
Our experiments are conducted under the environment of Python3.7 +Theano + Cuda10.0. The dimension of BERT is initialized as 128. The window size of BiLSTM is set to 3, the number of filters is 50, and the following dense layer has a hidden layer with 100 dimensions. For MDiconv, the dilation rates are 1, 2, and 4. The embedding and classification layers are standardized by dropout with a ratio of 0.5. For Multi-head attention, the head number is set to 8. The dimension of the MGLSTM is set as 256, and the cell unit number is set as 100. The learning rate is set to 0.001, and the batch size is 64. The mini-batch gradient descent is applied to optimize parameters.
Baselines
To evaluate the effectiveness of this new model, extensive contrast experiments are carried out with the following state-of-the-art triple extraction models:
(1) Novel tagging (Zheng et al., 2017): This model converts the joint extraction task to a sequential labeling problem by a tagging scheme where each token is assigned a unique tag denoting entity mentions and relation types simultaneously.
(2) CopyRE (Zeng et al., 2018): This RE model is an end-to-end neural model with a copy mechanism. The principle is that the relation is first extracted, and then the corresponding entity pair is extracted by a copy mechanism from the source texts.
(3) GraphRel (Fu and Ma, 2019): Based on GCNs, this end-to-end joint extraction model can predict relations between all word pairs. It constructs a complete word graph for each sentence accurately.
(4) CopyMTL (Zeng et al., 2020): Based on a multi-task learning framework equipped with a copy mechanism, CopyMTL is constructed to predict multi-token entities.
(5) HRL (Takanobu et al., 2019): This model applies a hierarchical reinforcement learning framework that decomposes the task into a high-level task for relation detection and a low-level task for entity extraction.
(6) RSAN (Yuan et al., 2020): The relation-aware attention mechanism is applied in RSAN to construct specific sentence representations for each relation. Then the corresponding head and tail entities are extracted by performing the sequence labeling.
(7) WDec (Nayak and Ng, 2020): A novel triples representation scheme is proposed, and the sequence-to-sequence mechanism is employed to produce the word sequences.
(8) TPLinker (Wang et al., 2020a): TPLinker treats the joint extraction task as a token pair linking problem to overcome the overlapping triple challenge. There are no interdependent stages; thus, the error accumulation is alleviated.
Performance Metrics
A relational triple is considered correct, where the two entities and the corresponding relation type are all correct. The standard Precision (Prec.), Recall (Rec.), and F1 scores are selected as the evaluation matrix of experiment results.
Experimental Results
Comparative Experiments
To show the effectiveness of the new model ROMGCJE in extracting triples, we first carry out the plenty of contrast experiments with some state-of-the-art methods on public datasets NYT and WebNLG. The test results are presented in Table 3, and the best experiment results are denoted by the bold numbers. It is easy to conclude that the performances of other existing models do not surpass our model on both two datasets. This new model achieves a 1.6 and 1.0% gain in F1 over the best method TPLinker on datasets NYT and WebNLG, respectively. The following points might contribute to the improvement of ROMGCJE: (1) The sufficient contextual information and long-range dependency captured by PTSRL and CWRL benefit NER and RC tasks greatly. (2) The MGLSTM can capture sufficient global information. The global embedding from MGLSTM contributes to the RC task, which excludes the error caused by predictions of the wrong relationship.
Performance on Normal, EPO, and SEO Sentences
To further explore the capability of ROMGCJE in extracting overlapping relations, extensive experiments are developed on the NYT dataset. Based on the triple overlapping degree, the NYT test set is divided into the following three sub-sets: Normal sentence set, SEO sentence set, and EPO sentence set. Figure 5 shows the F1 values of the different comparative models on NYT test data. The different color blocks denote the experiment results of the different models. The F1 score of ROMGCJE surpasses other models on normal, EPO, and SEO sentences. In particular, ROMGCJE shows more outcoming performance on the EPO sentences than other comparative models. There are three reasons for this performance as follows: (1) The new labeling scheme can assign different tags to a word, (2) the new model ROMGCJE can make separate predictions for the different relations, and (3) the two attention modules have some contributions to the prediction of EPO triples.
In the following parts, we conduct more ablation studies to verify the effectiveness of each module in ROMGCJE and further explore the reasons for the great improvement.
Ablation for Character Embedding
To evaluate the contribution of the character representation captured by BiLSTM in the PTSRL, some ablation tests are conducted on NYT and WebNLG datasets. The experiment results are shown in Table 4. “–Bidirectional long short-term memory neural network” means that the character embedding is not used in PTSRL, and the word representation learned by BERT is directly fed to CWRL. From Table 4, we can see that the introduction of character embedding promotes a 1.6% increment of F1. We attribute this phenomenon to that the character embedding includes the local feature of input text and greatly contributes to extracting the morphological information and dealing with the out-of-vocabulary problem.
Ablation for Attention Module
To demonstrate the contribution of the two attention modules, RBAGM and AJE, plenty of ablation tests are conducted on the NYT dataset. The experimental results are summarized in Table 4. “—RBAGM” denotes that the RBAGM module is not applied when extracting triples. Similarly, “–AJE” means that the AJE module is not applied in GCRE. As we can see from Table 4, the model's Precision drops significantly when RBAGM or AJE is deleted from ROMGCJE. We can conclude that the sentence representations fused with the fine-grained semantic relation feature greatly affect the joint extraction task.
Besides, we conduct the following two groups of experiments on NYT dataset: (1) Exploring the influence of the different distances between entities on ROMGCJE and (2) exploring the influence of sentence length on ROMGCJE. The experimental results are presented in Figures 6, 7, respectively, weighted by the F1 score. The entity distance is measured by the rule that the absolute character offset between the last character of the first occurring entity and the last character of the second-mentioned entity. From Figure 6 we can conclude that ROMGCJE significantly outperforms “–RBAGM” and “–AJE” across the different entity distances. The introduction of AJE and RBAGM makes the model ROMGCJE increase by 4.1 and 3.4% in F1 score, respectively, when the entity distance is more than 20 characters. It is easy to conclude that AJE and RBAGM contribute greatly to the triple extraction task. In the second group of experiments, we partition the NYT dataset into five groups based on the sentence length [(0–19), (20–29), (30–39), (40–49), (≥50)]. We analyze the performance of ROMGCJE, “–RBAGM,” and “–AJE” on these five subsets, as shown in Figure 7. We can observe a decline in the F1 score of these three models when sentences contain more words. However, the performance of ROMGCJE still outperforms that of “–RBAGM” and “–AJE.” Moreover, ROMGCJE outperforms “–RBAGM” and “–AJE” by 5.26 and 6.42% in the F1 score for triple extraction, respectively, when the sentence contains more than 40 words.

Figure 6. Comparison of ROMGCJE, “–RBAGM,” and “–AJE” under different entity distances on the NYT dataset.

Figure 7. Comparison of the ROMGCJE, “–RBAGM,” and “–AJE” under different sentence lengths on the NYT dataset.
We analyze the above results from the following two perspectives: (1) For AJE, it promotes the interaction between RC and NER tasks since it can pass entity information from NER task to RC task and collect RC task's feedback information by jointly updating all the parameters. (2) For RBAGM, it plays a significant role in capturing the global dependencies of the whole sentence, which greatly contributes to the prediction of EPO and SEO triples in long sentences. In EPO sentences, the relation is different while the entity pairs are the same; thus, these entities have more sufficient semantic information, which attracts more attention in decoding. Finally, it is more possible for these entities to be selected from the input sentence. Based on the above analysis, the model ROMGCJE has great advantages in extracting entities and relations from long sentences with EPO or SEO triples.
Ablation for CWRL Module
Plenty of ablation experiments are conducted on the NYT dataset to find the best combination of MHAttention and MDiconv. The experimental results are presented in Table 5. When the number of both MHAttention and MDiconv is set to 0, the PTSRL is directly applied to encode tokens without applying the max-pooling module. We can observe that the Prec., Rec., F1 scores drop of 13.8, 14.4, and 13.1%, respectively, indicating that CWRL is critical for improving model performance. Once the number of MHAttention or MDiconv increases, the model's performance improves to a different extent. This suggests that the contextual information obtained by CWRL can greatly assist ROMGCJE in jointly extracting entities and relations. After analyzing the prediction results of the model with different combinations of MHAttention and MDiconv, we conclude that the number of MHAttention and MDiconv is set to 2 and 2, respectively, making the ROMGCJE model achieve the best performance. Further, with the number of Mdiconv or MHAttention increasing from 1 to 5, the values of the evaluation matrix increase first but decrease later. Thus, we can obtain that more Mdiconv modules or MHAttention modules are not always better.
Ablation for MGLSTM
Plenty of ablation experiments are conducted on the NYT dataset to explore the effects of MGLSTM for NER in the decoder part. In this process, the MGLSTM is replaced by CRF and LSTM, respectively, and the final results are presented in Figure 8. Also, “replace MGLSTM with CRF” means that the MGLSTM is replaced by the CRF module in ROMGCJE; “replace MGLSTM with LSTM” refers to that LSTM replaces the MGLSTM. From Figure 8, we can see that the MGLSTM for NER makes the model achieve the best performance. Using CRF for NER leads to a reduction of 2.4% Precision in NYT. The reason is that there is a long distance among these relation tags, but CRF has difficulties in overcoming this problem. In contrast, the performance of LSTM is a little better than CRF. It only obtains a reduction of 1.8% Precision in NYT. This is because LSTM can build a long-range dependency to some extent. Besides the merits inherited from LSTM, MGLSTM can learn to represent information over multiple time scales, and the introduction of global embedding can reduce the damage caused by mispredictions made in the previous time steps. Therefore, the application of MGLSTM makes the model predict label sequences more accurately.
Ablation Study for MRC
To demonstrate the effectiveness of the MRC module in ROMGCJE, some experiments are developed under the circumstances that these entities are known. The comparison performance on the NYT dataset is presented in Figure 9. There are some different combinations as follows: (1) “replace MRC with BiLSTM” means that the MRC is replaced by BiLSTM (2) “replace MRC with Tree-LSTM” refers to that the MRC is replaced by Tree-LSTM (3) “replace MRC with Multisigmoid layer” represents that the MRC is replaced by Multisigmoid layer. From Figure 9, we can see that ROMGCJE achieves the best result in terms of triples extraction. The possible reason is that BiLSTM, Tree-LSTM, and Multisigmoid layer module have great difficulties in assigning multiple tags to one word, and thus they cannot deal with the overlapping relation problems. The above results demonstrate that the MRC module is very suitable for ROMGCJE. NYT contains much noise data, which indicates that ROMGCJE is robust.
Conclusions
This study proposes a novel joint extraction model ROMGCJE for overlapping relationships and entities. The introduction of CWRL enables ROMGCJE to capture rich global contextual information, build long-range dependency among words, and fully extract the semantics of the text. Besides, the global embeddings learned by MGLSTM boost the extraction of entities and reduce the error propagation from NER task to RC task. In addition, applying the attention mechanism contributes to the prediction of overlapping relation triples greatly. Comprehensive experiments prove that the proposed method achieves the state-of-the-art performance compared with other approaches.
Based on the model ROMGCJE, one self-learning KG can be developed, which has the ability to better organize, manage and understand the massive information on the Internet. In the future, this self-learning KG can be applied into many fields, such as search engines, intelligent question answering, intelligent recommendation, intelligent furniture, fault diagnosis, and so on. The ROMGCJE model will contribute to the research of the above-mentioned fields significantly.
Data Availability Statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author/s.
Author Contributions
HH: writing—original draft preparation, writing—reviewing and editing, conceptualization, validation, formal analysis, and methodology. JW: conceptualization, data curation, supervision, project administration, funding acquisition, and resources. XW: supervision, software, validation, formal analysis, investigation, reviewing and editing, and data curation. All authors contributed to the article and approved the submitted version.
Funding
This work was supported by the National Science and Technology Innovation 2030 of China Next-Generation Artificial Intelligence Major Project, Data-Driven Tripartite Collaborative Decision-Making and Optimization, under Grant 2018AAA0101801.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
The work described in this study was conducted at the Computer Integrated Manufacturing System Research Center, College of Electronics and Information Engineering, and Tongji University. The authors would like to thank Dr. Chen for supporting and encouraging.
References
Alam, F., and Islam, M. A. (2020). “A proposed model for bengali named entity recognition using maximum entropy markov model incorporated with rich linguistic feature set,” in ICCA 2020: International Conference on Computing Advancements, Vol. 64 (Assoc Computational Linguistics-ACL), 1–6 doi: 10.1145/3377049.3377117
Bekoulis, G., Deleu, J., Demeester, T., and Develder, C. (2018). Joint entity recognition and relation extraction as a multi-head selection problem. Expert Syst. Appl. 114, 34–35. doi: 10.1016/j.eswa.2018.07.032
Chen, J., Yuan, C., Wang, X., and Bai, Z. (2019). “MrMep: joint extraction of multiple relations and multiple entity pairs based on triplet attention,” in Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL) (Hong Kong: Association for Computational Linguistics), 593–602. doi: 10.18653/v1/K19-1055
Dai, D., Xiao, X., Lyu, Y., Dou, S., and Wang, H. (2019a). “Joint extraction of entities and overlapping relations using position-attentive sequence labeling,” in Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33 (Palo Alto, CA: AAAI Press), 6300–6308. doi: 10.1609/aaai.v33i01.33016300
Dai, Z., Wang, X., Ni, P., Li, Y., and Bai, X. (2019b). “Named entity recognition using BERT BiLSTM CRF for Chinese electronic health records,” in 12th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (Suzhou: IEEE), 1–5. doi: 10.1109/CISP-BMEI48845.2019.8965823
Dalvi, N., Kumar, R., and Soliman, M. (2011). Automatic wrappers for large scale web extraction. Proc. Vldb Endowment 4, 219–230. doi: 10.14778/1938545.1938547
Devlin, J., Chang, M. W., Lee, K., and Toutanova, K. (2018). BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Minneapolis, MN: Association for Computational Linguistics. doi: 10.48550/arXiv.1810.04805
Eberts, M., and Ulges, A. (2019). “Span-based joint entity and relation extraction with transformer pre-training,” in Proceedings of the 24th European Conference on Artificial Intelligence, Vol. 325 (European Assoc Artificial Intelligence, IOS PRESS), 2006–2013. doi: 10.3233/FAIA200321
Fu, T. J., and Ma, W. Y. (2019). “GraphRel: modeling text as relational graphs for joint entity and relation extraction,” in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Florence: Association for Computational Linguistics), 1409–1418. doi: 10.18653/v1/P19-1136
Gardent, C., Shimorina, A., Narayan, S., and Perez-Beltrachini, L. (2017). “Creating training corpora for NLG micro-planning,” in 55th Annual Meeting of the Association-for-Computational-Linguistics, Vol. 04 (Vancouver, BC: Association for Computational Linguistics), 179–188. doi: 10.18653/v1/P17-1017
Gong, L., Liu, X., Yang, X., Zhang, L., and Yang, R. (2019). CBLNER: a multi-models biomedical named entity recognition system based on machine learning. Intell. Comput. Theor. Applic. 11644, 51–60. doi: 10.1007/978-3-030-26969-2_5
Huang, Z., Wei, X., and Kai, Y. (2015). Bidirectional LSTM-CRF models for sequence tagging. arXiv [Preprint]. doi: 10.48550/arXiv.1508.01991
Li, X., Yin, F., Sun, Z., Li, X., Yuan, A., et al. (2019). “Entity-relation extraction as multi-turn question answering,” in 57th Annual Meeting of the Association-for-Computational-Linguistics, 1340–1350. doi: 10.18653/v1/P19-1129
Li, X. M., Luo, X. T., Dong, C. H., Yang, D. C., Luan, B. D., and He, Z. (2021). “TDEER: an efficient translating decoding schema for joint extraction of entities and relations,” in Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (Punta Cana: Association for Computational Linguistics), 8055–8064. doi: 10.18653/v1/2021.emnlp-main.635
Liu, B., Chiticariu, L., Chu, V., Jagadish, H. V., and Reiss, F. R. (2010). Automatic rule refinement for information extraction. Proc. VLDB Endowment 3, 588–597. doi: 10.14778/1920841.1920916
Liu, Y., Li, A., Huang, J., Zheng, X., and Wang, Z. (2019). “Joint extraction of entities and relations based on multi-label classification,” in 2019 IEEE Fourth International Conference on Data Science in Cyberspace (DSC), 106–111. doi: 10.1109/DSC.2019.00024
Liu, Z., Luo, X., and Wang, Z. (2021). Convergence analysis of single latent factor-dependent, non-negative, and multiplicative update-based nonnegative latent factor models. IEEE Trans. Neural Netw. Learn. Syst. 32, 1737–1749. doi: 10.1109/TNNLS.2020.2990990
Luo, X., Liu, W., Ma, M., and Wang, P. (2020). BiTT: Bidirectional Tree Tagging for Joint Extraction of Overlapping Entities and Relations. doi: 10.48550/arXiv.2008.13339
Miwa, M., and Bansal, M. (2016). “End-to-end relation extraction using lstms on sequences and tree structures,” in 54th Annual Meeting of the Association-for-Computational-Linguistics (Berlin: Assoc Computational Linguistics-Acl), 1105–1116. doi: 10.18653/v1/p16-1105
Nayak, T., and Ng, H. T. (2020). “Effective modeling of encoder-decoder architecture for joint entity and relation extraction,” in Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 34 (New York, NY: Assoc Advancement Artificial Intelligence), 8528–8535. doi: 10.1609/aaai.v34i05.6374
Och, F. J.. (2017). “Minimum error rate training in statistical machine translation. Association for Computational Linguistics,” in Proceedings of the 41st Annual Meeting on Association for Computational Linguistics (Sapporo: Association for Computational Linguistics), 160–167.
Read, J., Pfahringer, B., Holmes, G., and Frank, E. (2011). Classifier chains for multi-label classification. Mach. Learn. 85, 333–359. doi: 10.1007/s10994-011-5256-5
Ren, H., Cai, Y., Yiu, R., Lau, K., Leung, H., and Li, Q. (2022). Granularity-aware area prototypical network with bimargin loss for few shot relation classification. IEEE Trans. Knowl. Data Eng. doi: 10.1109/TKDE.2022.3147455
Riedel, S., Yao, L., and Mccallum, A. K. (2010). Modeling relations and their mentions without labeled text. Mach. Learn. Knowl. Discov. Databases 6323, 148–163. doi: 10.1007/978-3-642-15939-8_10
Salimans, T., and Kingma, D. P. (2016). “Weight normalization: a simple reparameterization to accelerate training of deep neural networks,” in 30th Conference on Neural Information Processing Systems (NIPS), Vol. 29 (Barcelona: Neural Information Processing Systems), 901–909.
Santos, C., Bing, X., and Zhou, B. (2015). Classifying relations by ranking with convolutional neural networks. Comput. Sci. 86, 132–137. doi: 10.3115/v1/P15-1061
Schuster, M., and Paliwal, K. K. (1997). Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 45:2673–2681. doi: 10.1109/78.650093
Skeppstedt, M., Kvist, M., Nilsson, G. H., and Dalianis, H. (2014). Automatic recognition of disorders, findings, pharmaceuticals and body structures from clinical text: an annotation and machine learning study. J. Biomed. Inform. 49, 148–158. doi: 10.1016/j.jbi.2014.01.012
Sun, C., Wu, Y., Lan, M., Sun, S., and Wu, K. (2018). “Extracting entities and relations with joint minimum risk training,” in Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (Belgium: Association for Computational Linguistics), 2256–2265. doi: 10.18653/v1/D18-1249
Takanobu, R., Zhang, T., Liu, J., and Huang, M. (2019). “A hierarchical framework for relation extraction with reinforcement learning,” in Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33 (Honolulu: Assoc Advancement Artificial Intelligence), 7072–7079. doi: 10.1609/aaai.v33i01.33017072
Tomori, S., Ninomiya, T., and Mori, S. (2016). “Domain specific named entity recognition referring to the real world by deep neural networks deep neural networks,” in Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Vol. 2 (Berlin: Assoc Computational Linguistics-ACL), 236–242. doi: 10.18653/v1/P16-2039
Vazquez, M., Krallinger, M., Leitner, F., and Valencia, A. (2011). Text mining for drugs and chemical compounds: methods, tools and applications. Mol. Inform. 30, 506–519. doi: 10.1002/minf.201100005
Wang, C., Li, A., Tu, H., Wang, Y., and Zhao, X. (2020b). “An advanced bert-based decomposition method for joint extraction of entities and relations,” in 2020 IEEE Fifth International Conference on Data Science in Cyberspace (DSC) (Hong Kong: IEEE), 82–88. doi: 10.1109/DSC50466.2020.00021
Wang, Y., Yu, B., Zhang, Y., Liu, T., and Sun, L. (2020a). “TPlinker: single-stage joint extraction of entities and relations through token pair linking,” in Proceedings of the 28th International Conference on Computational Linguistics (Barcelona: International Committee on Computational Linguistics), 1572–1582. doi: 10.18653/v1/2020.coling-main.138
Wei, H., Gao, M., Zhou, A., Chen, F., and Lu, M. (2019). Named entity recognition from biomedical texts using a fusion attention-based BILSTM-CRF. IEEE Access 7, 73627–73636. doi: 10.1109/ACCESS.2019.2920734
Wiseman, S., and Rush, A. M. (2016). “Sequence-to-sequence learning as beam-search optimization,” in Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing (Austin: Association for Computational Linguistics), 1296–1306. doi: 10.18653/v1/D16-1137
Wu, D., and Luo, X. (2020). Robust latent factor analysis for precise representation of high-dimensional and sparse data. IEEE/CAA J. Autom Sin. 8, 796–805. doi: 10.1109/JAS.2020.1003533
Yu, B., Zhang, Z., Shu, X., Wang, Y., Liu, T., Wang, B., et al. (2019). “Joint extraction of entities and relations based on a novel decomposition strategy,” in 24th European Conference on Artificial Intelligence (IOS Press, European Assoc Artificial Intelligence), 325, 2282–2289.
Yuan, Y.X, Zhou Pan, S., Zhu, Q., and Guo, L. (2020). “A relation-specific attention network for joint entity and relation extraction,” in Proceedings of the 7th Pacific Rim International Conference on Artificial Intelligence (Electr Network, IJCAI-INT Joint Conf Artif Intell), 4054–4060. doi: 10.24963/ijcai.2020/561
Zeng, D., Zhang, H., and Liu, Q. (2020). Copymtl: copy mechanism for joint extraction of entities and relations with multi-task learning. Proc. AAAI Conf. Artif. Intell. 34, 9507–9514. doi: 10.1609/aaai.v34i05.6495
Zeng, X., Zeng, D., He, S., Kang, L., and Zhao, J. (2018). “Extracting relational facts by an end-to-end neural model with copy mechanism,” in Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Vol. 1 (Palo Alto, CA: AAAI Press), 506–514. doi: 10.18653/v1/P18-1047
Zhang, R., Meng, F., Yong, Z., and Bing, L. (2018). Relation classification via recurrent neural network with attention and tensor layers. Big Data Mining Analyt. 1, 64–74. doi: 10.26599/BDMA.2018.9020022
Zhang, S., Sheng, Y., Gao, J., Chen, J., and Lin, S. (2019). A multi-domain named entity recognition method based on part-of-speech attention mechanism. Comput. Supported Cooper. Work Soc. Comput. 1042, 631–644. doi: 10.1007/978-981-15-1377-0_49
Zhang, Y., Zhong, V.D, Chen Angeli, G., and Manning, C. D. (2017). “Position-aware attention and supervised data improve slot filling,” in Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing (Copenhagen: Association for Computational Linguistics), 35–45. doi: 10.18653/v1/D17-1004
Zhao, T., Yan, Z., Cao, Y., and Li, Z. (2021). “A unified multi-task learning framework for joint extraction of entities and relations,” in Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 35 (Palo Alto, CA: AAAI Press), 14524–14531.
Zheng, S.F, Wang Bao, H., Hao, Y., Zhou, P., and Xu, B. (2017). “Joint extraction of entities and relations based on a novel tagging scheme,” in Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Vancouver, BC: Association for Computational Linguistics), 1227–1236. doi: 10.18653/v1/P17-1113
Keywords: joint extraction of entities and relations, multi-label classification, relation extraction, entity recognition, overlapping triples
Citation: Han H, Wang J and Wang X (2022) A Relation-Oriented Model With Global Context Information for Joint Extraction of Overlapping Relations and Entities. Front. Neurorobot. 16:914705. doi: 10.3389/fnbot.2022.914705
Received: 07 April 2022; Accepted: 19 May 2022;
Published: 04 July 2022.
Edited by:
Xin Luo, Chongqing Institute of Green and Intelligent Technology (CAS), ChinaReviewed by:
Suna Bensch, Umeå University, SwedenBin Zhi Li, Chongqing Institute of Green and Intelligent Technology (CAS), China
Copyright © 2022 Han, Wang and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Huihui Han, aGFuaHVpaHVpQHRvbmdqaS5lZHUuY24=; Xiaowen Wang, d2FuZ3h3MjBAbWFpbHMudHNpbmdodWEuZWR1LmNu