Next Article in Journal
DanceTrend: An Integration Framework of Video-Based Body Action Recognition and Color Space Features for Dance Popularity Prediction
Next Article in Special Issue
A Neighborhood-Similarity-Based Imputation Algorithm for Healthcare Data Sets: A Comparative Study
Previous Article in Journal
Guaranteeing Zero Secrecy Outage in Relaying Systems under Eavesdropper’s Arbitrary Location and Unlimited Number of Antennas
Previous Article in Special Issue
Resolving Agent Conflicts Using Enhanced Uncertainty Modeling Tools for Intelligent Decision Making
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Joint Overlapping Event Extraction Model via Role Pre-Judgment with Trigger and Context Embeddings

1
School of Computer and Information Technology, Shanxi University, Taiyuan 030006, China
2
Key Laboratory of Computational Intelligence and Chinese Information Processing of Ministry of Education, Shanxi University, Taiyuan 030006, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(22), 4688; https://doi.org/10.3390/electronics12224688
Submission received: 29 October 2023 / Revised: 16 November 2023 / Accepted: 16 November 2023 / Published: 18 November 2023
(This article belongs to the Special Issue Advances in Intelligent Data Analysis and Its Applications, Volume II)

Abstract

:
The objective of event extraction is to recognize event triggers and event categories within unstructured text and produce structured event arguments. However, there is a common phenomenon of triggers and arguments of different event types in a sentence that may be the same word elements, which poses new challenges to this task. In this article, a joint learning framework for overlapping event extraction (ROPEE) is proposed. In this framework, a role pre-judgment module is devised prior to argument extraction. It conducts role pre-judgment by leveraging the correlation between event types and roles, as well as trigger embeddings. Experiments on the FewFC show that the proposed model outperforms other baseline models in terms of Trigger Classification, Argument Identification, and Argument Classification by 0.4%, 0.9%, and 0.6%. In scenarios of trigger overlap and argument overlap, the proposed model outperforms the baseline models in terms of Argument Identification and Argument Classification by 0.9%, 1.2%, 0.7%, and 0.6%, respectively, indicating the effectiveness of ROPEE in solving overlapping events.

1. Introduction

Event extraction (EE) is a challenging task in natural language understanding that plays a crucial role [1,2,3,4]. EE is dedicated to extracting event information occurring in the real world from text, classifying them into predefined event types, identifying the trigger and event participants, etc. [5]. Event extraction can be widely used in information retrieval and summarization, knowledge graph construction [6], intelligence analysis [7], and other fields. EE usually includes four subtasks, i.e., trigger identification (TI), Trigger Classification (TC), Argument Identification (AI), and Argument Classification (AC) [8,9,10].
The objective of event extraction is to identify events and arguments from the text. The event extraction task is formally defined as follows:
Given an input sentence x = {x1, x2, …, xN} consisting of N words, an event-type collection ( C ), and an argument role collection ( ). Let x represent the gold set. x includes all event types ( C x C ) in the sentence, all triggers ( T x , c ) in each event type ( c C x ), and different argument roles under each event type. a r A x is the argument corresponding to r .
However, a sentence may contain multiple events, and the arguments and triggers of these events have complex overlapping phenomena. We summarize this into three situations: (1) Different event types may be triggered by the same trigger. The trigger “减持” (reduced its holdings) marked in red in Figure 1 triggers both the shareholding reduction event and the share equity transfer event. (2) The same argument plays different roles in different event types. In Figure 1, “大族激光” (Han’s Laser) plays the role of obj in the “股份股权转让” (share equity transfer) event, and also plays the target-company role in the share equity transfer, which is triggered by “减持” (reduced its holdings). (3) The same argument plays the same role in different events of the same event type. In Figure 1, the “股份股权转让” (share equity transfer) event appears twice in the sentence, both occurring in “10月” (October). In FewFC (a Chinese financial event extraction dataset) [11], about 13.5% of the sentences have trigger overlap problems, and 21.7% of the sentences have argument overlap problems. The same event type appears repeatedly in 8.3% of the sentences.
Most previous studies partially addressed overlapping issues and did not cover all the above situations. In 2019, Yang et al. [12] utilized a staged pipeline approach for event trigger and argument extraction. Nevertheless, this method overlooked the challenge of trigger overlap. In 2020, Xu et al. [13] used a joint extraction framework to solve the role overlapping problem. Xu defined event relationship triples to represent the relationship between triggers, arguments, and roles, thereby converting the argument classification problem into a relationship extraction problem. In 2020, Huang et al. [14] proposed using a hierarchical knowledge structure graph containing conceptual and semantic reasoning paths to represent knowledge. They employed GEANet to encode intricate knowledge, addressing the issue of trigger extraction in nested structures within the biomedical domain [14]. In 2022, Zhang et al. [15] designed a two-stage pipeline model in which the trigger is identified using a sequence annotation approach, and overlapping arguments are identified through multiple sets of role binary classification networks. In 2023, Yang et al. [16] used a multi-task learning model to extract entity relationships and events, in which a multi-label classification method is used to settle the overlapping role problem shared by these two tasks.
The contributions of this paper are summarized as follows:
(1)
A role pre-judgment module is proposed to predict roles based on the correspondence between event types and roles, text embeddings, and trigger embeddings, which can significantly improve the recall rate of each subtask and provide a basis for extracting overlapping arguments.
(2)
ROPEE adopts a joint learning framework, and the designed loss function includes the losses of four modules, event-type detection, trigger extraction, role pre-judgment, and argument extraction, so as to effectively learn the interactive relationship between modules during training. Thus, error propagation issues can be alleviated in the prediction stage.
(3)
ROPEE outperforms the baseline model by 0.4%, 0.9%, and 0.6% in terms of F1 over TC, AI, and AC on the FewFC dataset. For sentences with overlapping triggers, ROPEE outperforms the baseline model by 0.9% and 1.2% in terms of F1 over AI and AC, respectively. In the case of overlapping arguments, ROPEE demonstrates superior performance compared to the baseline model, with improvements of 0.7% and 0.6%. This highlights the effectiveness of our suggested approach in managing overlapping occurrences of event phenomena.
The remainder of this paper is organized as follows: In Section 2, related studies are given. In Section 3, the details of the ROPEE model are introduced. Comparative experiments are performed and experimental results are analyzed in Section 4. Section 5 concludes this work.

2. Related Studies

Event extraction is one of the most challenging tasks in information extraction research [17]. Existing paradigms related to event extraction include pipeline methods and joint learning methods [18].
The pipeline-based method handles these four subtasks of EE separately. Each subtask has its own objective function and loss. In 2015, Chen et al. [8] developed a dynamic multi-pooling convolutional neural network (DMCNN). This network utilizes a dynamic multi-pooling layer based on event triggers and arguments to retain essential information, combining both the sentence-level and lexical-level details from the raw text without the need for extensive preprocessing. Most deep learning supervised methods for event extraction require lots of labeled data for training. Annotating large amounts of data is very laborious and hard to get. To gain more insights from limited training data, Yang et al. [12] combined the extraction model and event generation method in 2019 and improved the performance of the argument extractor through a weighted loss function based on various role importance. The above two methods cannot explicitly model the semantics between events and roles, nor can they capture the interaction between them. In 2020, Li et al. [19] devised a multi-stage QA framework to represent event extraction as reading comprehension issues and captured the dynamic connection between each subtask by integrating previous answers into questions. The generative event extraction model proposed by Paolini et al. [20] in 2021 solves the encoding problem of label semantics and other weak supervision signals in a pipeline manner and can improve the performance in few-sample scenarios. Since the loss function in the pipeline-based method is calculated after the argument extraction, error propagation problems may occur.
Joint learning methods integrate the loss in both the trigger extraction stage and argument extraction stage into the final loss function, treating triggers and arguments equally, and the two can mutually promote each other’s extraction effects [18]. In 2021, Sheng et al. [9] first covered all event overlap issues through a unified framework with a cascading decoder to perform TC, TI, and argument extraction in sequence, and F1 reached 71.4% on the FewFC dataset. In order to further extract inter-word relationships in overlapping sentences in parallel, Cao et al. [10] proposed a single-stage framework based on inter-word relationships by jointly extracting the intra-word and cross-word pair relationships of triggers and arguments. The above two methods focus on the event extraction task itself and do not introduce additional information or other tasks of joint information extraction. In 2022, Hsu et al. [21] converted event extraction into a conditional generation problem, and extracted triggers and arguments end-to-end through additional prompts and weak supervision information. In 2022, Van Nguyen et al. [22] used an edge weighting mechanism to learn the dependency graph between task instances and jointly complete the information extraction task. In addition to introducing additional prompt information in document-level event extraction, remote dependencies can also be used to improve extraction performance. In 2023, Liu et al. [23] proposed a chain reasoning paradigm for document-level event argument extraction, which represented argument queries by constructing first-order logic rules and T-Norm fuzzy logic, which is used for end-to-end learning. We propose a joint overlapping event extraction model ROPEE for the event overlapping phenomenon. It uses the correspondence between event types and roles and trigger embeddings to predict roles, which not only effectively alleviates error propagation, but also further improves the accuracy of event extraction.

3. ROPEE Model

The overall framework of ROPEE is illustrated in Figure 2. ROPEE includes four modules: event detection, trigger identification, role pre-judgment, and argument extraction. Specifically, type detection predicts potential event types and extracts overlapping triggers by calculating the similarity between sentence representations and event-type embeddings. The role pre-judgment module comprehensively considers text embeddings and trigger embeddings, pre-judgment roles based on the correspondence table between event types and roles, which can assist in the extraction of overlapping arguments. Trigger extraction and argument extraction are based on text representation that incorporates specific event types of information and specific role information, and binary classifiers are adopted to predict the starting and ending positions of triggers or arguments. To minimize error propagation, all modules are jointly learned during training.

3.1. Encoder

BERT [24] is utilized as an encoder. Sentence x = {x1, x2, …, xN} treats each Chinese character (xi) as a token and is fed into the bert-base-Chinese module. The embedding of the sentence is obtained by H = BERT (x1, x2, …, xN) = {h1, h2, …, hN} N × d , where d is the dimension of the embeddings.

3.2. Event Detection Decoder

The event detection decoder is shown in the upper left corner of Figure 2, which is used to predict potential event types in sentences by calculating the correlation between sentence representations that imply type features and event-type embeddings. Specifically, event-type embeddings can be denoted by a randomly initialized matrix (C | C | × d ). We apply rel to calculate the relevance between each token embedding (hi) and the potential event type (cC), see Equation (1), and then a sentence representation (sc) adaptive to the event type is obtained, see Equation (2). Thus, the similarity probability between sc and c is generated using a normalization σ operation for each event, see Equation (3).
r e l ( c , h i ) = v T × t a n h ( W r e l 1 [ c ; h i ; | c h i | ; c h i ] )
s c = i = 1 N e x p ( r e l ( c , h i ) ) j = 1 N e x p ( r e l ( c , h j ) ) h i    
c ^ = σ ( r e l ( c , s c ) )
where W r e l 1 4 d × 4 d and v 4 d × 1 are parameters of relevance calculation. | · | denotes the element-wise subtraction operation. denotes the element-wise multiplication. [ · ; · ] represents the concatenating operation. σ represents the sigmoid function. Types satisfying c ^ > ξ 1 are selected as potential event types, where ξ 1 is a threshold hyperparameter between 0 and 1. All potential event types hidden in sentence x constitute the set of event types C x . The decoder can learn the parameters θ t d = { W r e l 1 , v , C } .

3.3. Trigger Identification Decoder

A large number of experiments demonstrate that trigger information can enhance the ability of argument extraction. The trigger identification decoder is used to identify triggers according to a specific event type ( c C x ). The decoder includes a conditional layer normalization (CLN) [25], a self-attention layer [26], and a binary trigger tagging classifier pair.
CLN fuses the two features and filters out unnecessary information. Here, the event-type information is encoded into the token representation, and the event-typed token representation ( g i c ) is obtained:
g i c = CLN ( c , h i ) = γ c ( h i μ σ ) + β c
where type embedding c is used as the condition for γ c = W γ c + b γ and β c = W β c + b β in CLN. μ and σ are regarded as the average and deviation of hi:
μ = 1 d k = 1 d h i k ,       σ = 1 d k = 1 d ( h i k μ ) 2
where hik represents the k-th dimension of hi.
In order to fully consider the contextual connection in the sentence, a self-attention layer [26] is used on the event-typed token representation:
Z c = SelfAttention ( G c )
where G c = { g 1 c , g 2 c , , g N c } , G c N × d .
For each token, the binary classifier pair can mark the beginning and end position of a trigger span:
t ^ i s c = p ( t s | x i , c ) = σ ( w t s T z i c + b t s ) t ^ i e c = p ( t e | x i , c ) = σ ( w t e T z i c + b t e )
where z i c stands for the i-th token embedding in Z c . We select the token satisfying t ^ i s c > ξ 2 as the start position, and the one satisfying t ^ i e c > ξ 3 as the end. ξ 2 and ξ 3 are threshold hyperparameters. To acquire trigger t, each starting position is enumerated and the nearest subsequent ending position is searched in the sentence. A token span from the start position to the end constitutes a complete trigger. The corresponding triggers are extracted at different stages according to the potential event type. Thus, the trigger overlapping problem can be solved naturally. The set T c , x contains all predicted triggers (t) under event type (c) in sentence (x). θ t e is used to denote all parameters in the trigger identification decoder module.

3.4. Role Pre-Judgment Decoder

Since not all roles appear in a sentence under a specific event type, we designed a role pre-judgment decoder. Based on the predicted event type, it predicts the roles appearing in the sentence based on the corresponding list of event types and roles, providing a basis for extracting overlapping arguments. The decoder consists of three parts: the conditional fusion layer, the self-attention layer, and a role similarity detection function.
In order to obtain richer semantic information, we use CLN to fully integrate trigger embeddings and token representation with event-type knowledge to obtain a new token representation ( g i c t ), see Equation (8). Here, trigger embedding t is calculated by the average pooling of token embeddings in the trigger span.
g i c t = CLN ( g i c , t )
The self-attention layer then reinforces the contextual relationships and the sentence representation Z c t is obtained:
Z c t = SelfAttention ( G c t )
where G c t = { g 1 c t , g 2 c t , , g N c t } , G c t N × d .
The role similarity detection function predicts potential event-type-specific roles in sentences by calculating the correlation between role embeddings and sentence representations fused with role feature information. Specifically, a randomly initialized matrix R | | × d is used as role embeddings. We apply rel to calculate the relevance between each token embedding ( z i c t ) and the potential role (rR), see Equation (10), and then a sentence representation (sr) adaptive to the role is obtained, see Equation (11). Thus, the similarity probability between sr and r is generated, see Equation (12), based on which normalization operation is performed to obtain the predicted probabilities of all roles under a specific event type.
r e l ( r , z i c t ) = v T × t a n h ( W r e l 2 [ r ; z i c t ; | r z i c t | ; r z i c t ] )
s r = i = 1 N e x p ( r e l ( r , z i c t ) ) j = 1 N e x p ( r e l ( r , z j c t ) ) z i c t  
r ^ c t = p ( r | x i , c , t ) = σ ( r e l ( r , s r ) )
Select the role that satisfies r ^ c t > ξ 4 as the potential role type, and ξ 4 is the threshold. The role type set t , c , x contains all potential roles whose trigger is t under event type c in sentence x. θ r e is used to denote all parameters in the role pre-judgment decoder module.

3.5. Argument Extraction Decoder

An argument extraction decoder is composed of a positional embedding layer (PEL) and role-aware binary classifier pairs for argument tagging.
The relative position of a token to the trigger in the text is beneficial for argument extraction [9,27]. Here, relative position embeddings [8] imply the relative distance information between the current token and the trigger boundary token. Relative positional embeddings are incorporated into the sentence representation ( Z c t ) using a concatenation operation:
Z c t = [ Z c t ; P ]
where P N × d p is the relative position embeddings, and dp is the dimension of position embeddings.
For each token, a binary classifier sequence is employed to mark the boundary position of an argument under r t , c , x :
a ^ i s c t r = p ( a r s | x i , c , t , r ) = r ^ c t σ ( w r s T z i c t + b r s ) a ^ i e c t r = p ( a r e | x i , c , t , r ) = r ^ c t σ ( w r e T z i c t + b r e )
where z i c t represents the i-th token in Zct. For each role (r), select the token that satisfies a ^ i s c t r > ξ 5 as the starting position and the one that satisfies a ^ i e c t r > ξ 6 as the end. ξ 5 , ξ 6 [ 0 ,   1 ] are thresholds. In order to extract the boundary of argument (ar) with role (r), all starting positions are enumerated and the nearest subsequent ending position is searched in the sentence. Tokens between the starting and the ending position constitute a complete argument. In this way, only arguments of a specific role (r) under a specific trigger (t) and specific event type (c) in a sentence are extracted at each prediction stage. Thereby, the argument overlapping problem can be solved naturally. All candidate arguments (ar) form a set A r , t , c , x , and θ a e denotes the set of all parameters of the PEL and argument classifier.

3.6. Model Training

The loss of four modules is integrated during the training process, so the total loss function is designed as follows:
Loss all = x D [ c C x l o g p θ 1 ( c | x ) + t T x , c l o g p θ 2 ( t | x , c ) + r x , c , t l o g p θ 3 ( r | x , c , t ) + a r A x , c , t , r l o g p θ 4 ( a r | x , c , t , r ) ]
where Θ { θ 1 , θ 2 , θ 3 , θ 4 } . The first two subtasks, p θ 1 ( c | x ) and p θ 2 ( t | x , c ) , are adopted from [9]. We decomposed the argument extraction loss to p θ 3 ( r | x , c , t ) and p θ 4 ( a r | x , c , t , r ) and formulated it as:
p θ 1 ( c | x ) = ( c ^ ) c ¯ ( 1 c ^ ) ( 1 c ¯ ) p θ 2 ( t | x , c ) = z { s , e } i = 1 N ( t ^ i z c ) t ¯ i z c ( 1 t ^ i z c ) ( 1 t ¯ i z c ) p θ 3 ( r | x , c , t ) = ( r ^ c t ) r ¯ c t ( 1 r ^ c t ) ( 1 r ¯ c t ) p θ 4 ( a r | x , c , t , r ) = z { s , e } i = 1 N ( a ^ i z c t r ) a ¯ i z c t r ( 1 a ^ i z c t r ) ( 1 a ¯ i z c t r )
where c ^ , t ^ i s c , t ^ i e c , r ^ c t , a ^ i s c t r , and a ^ i e c t r are the predicted probabilities of the event type, starting and ending positions of triggers, role types, and starting and ending positions of arguments, respectively, which can be calculated according to Formulas (3), (7), (12), and (14). c ¯ , t ¯ i s c , t ¯ i e c , r ¯ c t , a ¯ i s c t r , and a ¯ i e c t r are real labels in the training data. θ 1 { θ b e r t , θ t d } , θ 2 { θ b e r t , θ t e } , θ 3 { θ b e r t , θ r e } , θ 4 { θ b e r t , θ a e } , and θ b e r t , θ t d , θ t e , θ r e , θ a e denote the parameters from BERT, event detection, trigger identification, role pre-judgment, and argument extraction, respectively. We choose Adam [28] over shuffled mini-batches to minimize Lossall.

4. Experiments and Analysis

4.1. Datasets

We use the FewFC dataset [11] to conduct comparative experiments. The reason for choosing FewFC is that other datasets do not completely cover the three overlapping situations mentioned in Section 1 like FewFC. For example, only 10% of events in the mainstream ACE2005 dataset have overlapping arguments, and there are no samples with overlapping triggers [12]. FewFC is a benchmark dataset in the Chinese financial field extracted for overlapping events, in which a total of 10 event types and 18 roles are annotated, and about 22% of the sentences contain overlapping events. Regardless of whether the event types are the same, the test set in FewFC contains 168 samples with overlapping triggers and 209 samples with overlapping arguments. Note that there is an intersection between overlapping samples. The dataset is split into training, validation, and testing by 80%, 10%, and 10%. See Table 1 for more detail.

4.2. Implementation Details

We choose PyTorch for code implementation, and NVIDIA A100-PCIE-40GB for training. ROPEE uses the bert-base-Chinese model with a starting learning rate of 2 × 10–5, a decoder learning rate of 1 × 10–4, and a decoder dropout rate of 0.3. The batch size is eight. The hidden layer size (d) is 768, the size of the position embeddings (dp) is 64, and the epochs for training are 20. All hyperparameters are turned on for the validation set. The event type embeddings and role embeddings are trained from scratch with random initialization.

4.3. Evaluation Metric

The evaluation metric includes four parts [8,9,10]: (1) Trigger Identification (TI): The trigger is considered correctly recognized if the predicted trigger span aligns with the ground truth label. (2) Trigger Classification* (TC): The trigger is deemed correctly classified when it is both accurately identified and assigned to the right event type. (3) Argument Identification (AI): The argument is considered correctly identified if the event type is accurately recognized and the predicted argument span aligns with the gold span. (4) Argument Classification (AC): The argument is considered correctly classified when it is both accurately identified and the predicted role matches the gold role. Each of these parts is evaluated by three metrics: micro precision (P), micro recall (R), and micro F1-measure (F1). The specific formula is as follows:
Precision m i c r o = i = 1 n TP i i = 1 n TP i + i = 1 n FP i Recall m i c r o = i = 1 n TP i i = 1 n TP i + i = 1 n FN i F 1 m i c r o = 2 · Precision m i c r o · Recall m i c r o Precision m i c r o + Recall m i c r o
where Precisionmicro and Recallmicro represent the average precision and recall across all categories.

4.4. Baselines

The following baseline models are chosen to compare with ROPEE over the FewFC dataset:
(1)
BERT-softmax [24]: It uses BERT to obtain the feature representation of words for the classification of both trigger and event arguments.
(2)
BERT-CRF [29]: It uses a CRF module based on BERT to catch the transfer rules between adjacent tags.
(3)
BERT-CRF-joint: It extends the classic BIO labeling scheme by merging tags of event types and roles for sequence annotation, such as BIO-type roles [30].
These methods convert the EE task into a sequence labeling task by attaching a label to each token. The flattened sequence labeling approach cannot address the overlapping problem due to label conflicts.
(4)
PLMEE [12]: It extracts triggers and arguments in a pipeline manner, and alleviates the argument overlapping issue by extracting role-aware arguments.
(5)
MQAEE: It is extended based on Li et al. [15]. It first predicts overlapping triggers through question and answer, and then predicts overlapping arguments based on typed triggers.
(6)
CasEE [9]: It performs all four subtasks sequentially with a cascade decoder based on the specific previous predictions.
The above three methods are multi-stage methods for overlapping event extraction.

4.5. Main Results

Table 2 shows experimental results comparing ROPEE with baseline models over FewFC. It can be revealed from the table below.
(1)
In contrast to the flattened sequence labeling methods, ROPEE achieves superior recall and F1 scores. Specifically, ROPEE outperforms BERT-CRF-joint by 15.7% and 4.9% on recall and the F1 score of AC. ROPEE also achieves significantly better than the sequence labeling method on recall because the sequence labeling method can only solve the flat event extraction problem, which can cause label conflicts. This shows that ROPEE can effectively solve the problem of overlapping event extraction.
(2)
Compared with the multi-stage methods for overlapping event extraction, the F1 scores of ROPEE in AI and AC are greater than that of CasEE by 0.9% and 0.6%, respectively. We believe that the role pre-judgment decoder in the model provides good help for argument extraction. In particular, ROPEE outperforms CasEE on the recall score of all four subtasks, especially AI and AC by 5.5% and 5.4%, respectively. This shows that the ROPEE model can better recall arguments that match the role type in the training of role pre-judgment. Overall, ROPEE outperforms all multi-stage methods for overlapping event extraction.
(3)
We also conducted comparative experiments on the large language model ChatGLM2-6B [31,32]. Some parameters of ChatGLM were fine-tuned using the P-Tuning-v2 method, but the final result was not ideal. ChatGLM is more accurate in extracting core arguments and triggers, but the span position it extracts is seriously inconsistent with the original text, indicating that ChatGLM fails to understand the boundary meaning represented by span.

4.6. Results of Overlapped EE

Table 3 shows the comparative performance of ROPEE and CasEE in two overlapping situations. For sentences with overlapping triggers, ROPEE achieves improvements of 0.9% and 1.2% over CasEE on the F1 score of AI and AC, respectively. For overlapping arguments, ROPEE outperforms CasEE in all four subtasks. Experimental results illustrate the superiority of ROPEE in solving overlapping problems, and we argue that the role pre-judgment decoder predicts potential roles and promotes the performance of extraction in overlapping arguments.

4.7. Ablation Studies

To verify the rationality of the role pre-judgment decoder, we conducted ablation experiments. In Table 4, ROP-text_emb only uses the text representation containing event-type embeddings as the predicted role type. type_emb denotes models that pre-judge roles using only event-type embeddings. As shown from this table, the type_emb used in the CasEE model performs worse than ROPEE on the F1 of TC, AI, and AC. It can be seen that event-type embeddings alone are not as efficient as the text embeddings containing event-type embeddings when used as the input for the role pre-judgment decoder, e.g., ROP-text_emb. ROPEE performs role pre-judgment by fusing the information of text representation and trigger embeddings, which outperforms ROP-text_emb in the F1 score of all four subtasks. It can be seen that trigger embeddings serve as supplementary information, allowing the model to grasp the knowledge hidden in the text.
In order to check the effectiveness of the role pre-judgment decoder, we adopt two training strategies while retaining the role pre-judgment decoder. The experimental results are given in Table 5. We use the correct role type for training and validation, such as Role-x (x can be 1, 2, 3, 4, or 5, representing the role pre-judgment decoder in the loss function after the correct role is input to the model weight value). With hyperparameters unchanged, the Role-1 model converged quickly. It stopped iterating at 11 epochs, and its effects on four subtasks were not as good as ROPEE. To enable the model to learn more fully and pay attention to the learning of role types, we gradually increase the weight of the loss function in the role prediction part with the other hyperparameters remaining unchanged. An obvious trend of first increasing and then decreasing can be seen from Role-x in Table 5, and the best performance is achieved in Role-3, which is only 0.3% lower than ROPEE in both of AI and AC. It can be seen that the training strategy of using the correct role type requires adjusting parameters on the original basis to enable the model to focus on role-type learning. However, the performance of this training strategy is not as good as ROPEE. We argue that the correct role type information can greatly reduce the difficulty of the task. The rapid convergence of the model during training leads to insufficient learning and over-fitting.
Table 6 shows the comparative results between a model with trigger extraction and one without trigger extraction. In the stage of data annotation, if no triggers are annotated, the event extraction task is completed using only event types, roles, and arguments. Since AI and AC tasks need only event-type information, we use these two to measure the performance of ROPEE. According to Table 6, 20 epochs are also used for training. When the correct triggers are not provided, the F1 score is decreased by 1.8% and 1.5% respectively, in AI and AC. Due to the increased difficulty in model training, we increased the number of training epochs to 50, and the result was increased by 0.2% in the F1 of AI, in contrast with 20 epoch w/o trigger. Thus, the lack of trigger annotation information will result in the inability to obtain the position embeddings of triggers during argument extraction, which also affects the model performance. Despite this, the model without trigger information achieves improvements than that of PLMEE and MQAEE, both of which are multi-stage methods with triggers involved.

4.8. Case Analysis

Figure 3 shows a test case analysis on CasEE and ROPEE, two events both triggered by “reduce”. For the “减持” (Reduction) event, both CasEE and ROPEE predict “Great Wall Film and Television” as the role of “target-company”, and there is no such argument in the standard answer. This shows that these two models lack the ability to judge the semantic boundaries of events. For the “股份股权转让” (share equity transfer) event, CasEE failed to extract the remote argument “Great Wall Group” as the role sub, while ROPEE improved the recall of AC by adding a role pre-judgment decoder, and, thus, was able to correctly detect the role of sub.

5. Conclusions

We design a joint learning framework ROPEE for overlapping event extraction, in which the event detection decoder can identify potential event types and help extract overlapping triggers. Based on text embeddings and trigger embeddings, a role pre-judgment module is proposed to predict roles based on the corresponding relationship between event types and roles, thereby enhancing the extraction of overlapping arguments. ROPEE has a certain effectiveness in addressing the issue of trigger overlap in EE, as well as argument overlap. Unlike other pipeline models, we integrate the tasks of four modules in the loss optimization layer, avoiding the traditional problem of error propagation existing in other pipeline methods. On the FewFC dataset, which was compared with other flattened sequence labeling methods (such as BERT-softmax, BERT-CRF, and BERT-CRF-joint), ROPEE achieves excellent recall and F1 scores on all subtasks; compared with multi-stage methods for overlapping event extraction (such as PLMEE, MQAEE, and CasEE), and ROPEE is superior to the above methods in terms of F1 scores of TC, AI, and AC. The above results show the superiority of ROPEE in overlapping event extraction. The ablation experimental results show that our model can also be used in different training strategies and other task scenarios or datasets without trigger labeling. In the future, we plan to build one more overlapping datasets based on specific application scenarios to verify our model’s performance. In addition, since trigger tagging requires additional manpower and material resources, the absence of trigger tagging will reduce the accuracy of the model. The scenario without trigger annotation is general (models should automatically find the core arguments in the sentence), and we plan to improve the model performance in this scenario using position embeddings relative to the core arguments. Finally, we implement joint learning by designing a joint loss of four modules in this article, and we plan to develop different joint strategies to strengthen the interaction between each subtask.

Author Contributions

Conceptualization, Q.C. and K.Y.; methodology, K.Y. and Q.C.; software, K.Y.; investigation, S.W.; resources, X.G.; writing—original draft preparation, K.Y.; writing—review and editing, J.Z.; supervision, X.G.; project administration, J.L.; funding acquisition, S.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially supported by National Key Research and Development Program of China (grant NO. 2022QY0300-01), National Natural Science Foundation of China (grant NO. 62076158), Natural Science Foundation of Shanxi Province of China (grant NOs. 202203021221021, 20210302123468 and 202203021221001), and CCF-Zhipu AI Large Model Fund (grant NO. 202310).

Data Availability Statement

The dataset used for this work is the FewFC dataset published by Zhou et al. [11]. It can be found at https://github.com/TimeBurningFish/FewFC (accessed on 17 November 2023). The codes are at https://github.com/yang-4074/ROPEE (accessed on 17 November 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Miwa, M.; Bansal, M. End-to-end relation extraction using lstms on sequences and tree structures. arXiv 2016, arXiv:1601.00770. [Google Scholar]
  2. Katiyar, A.; Cardie, C. Investigating lstms for joint extraction of opinion entities and relations. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 7–12 August 2016; pp. 919–929. [Google Scholar]
  3. Fei, H.; Zhang, M.; Ji, D. Cross-lingual semantic role labeling with high-quality translated training corpus. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 7014–7026. [Google Scholar]
  4. Li, J.; Xu, K.; Li, F.; Fei, H.; Ren, Y.; Ji, D. MRN: A locally and globally mention-based reasoning network for document-level relation extraction. In Proceedings of the Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, Online, 1–6 August 2021; pp. 1359–1370. [Google Scholar]
  5. Liu, Z.; Li, Y.; Zhang, Y.; Weng, Y.; Yang, K.; Wang, C. Effective Event Extraction Method via Enhanced Graph Convolutional Network Indication with Hierarchical Argument Selection Strategy. Electronics 2023, 12, 2981. [Google Scholar] [CrossRef]
  6. Bosselut, A.; Le Bras, R.; Choi, Y. Dynamic neuro-symbolic knowledge graph construction for zero-shot commonsense question answering. In Proceedings of the 35th AAAI conference on Artificial Intelligence, Online, 2–9 February 2021; pp. 4923–4931. [Google Scholar]
  7. Xiang, G.; Shi, C.; Zhang, Y. An APT Event Extraction Method Based on BERT-BiGRU-CRF for APT Attack Detection. Electronics 2023, 12, 3349. [Google Scholar] [CrossRef]
  8. Chen, Y.; Xu, L.; Liu, K.; Zeng, D.; Zhao, J. Event extraction via dynamic multi-pooling convolutional neural networks. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China, 26–31 July 2015; pp. 167–176. [Google Scholar]
  9. Sheng, J.; Guo, S.; Yu, B.; Li, Q.; Hei, Y.; Wang, L.; Liu, T.; Xu, H. CasEE: A joint learning framework with cascade decoding for overlapping event extraction. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021; Association for Computational Linguistics: Cedarville, OH, USA, 2021; pp. 164–174. [Google Scholar]
  10. Cao, H.; Li, J.; Su, F.; Li, F.; Fei, H.; Wu, S.; Li, B.; Zhao, L.; Ji, D. OneEE: A one-stage framework for fast overlapping and nested event extraction. In Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea, 12–17 October 2022; pp. 1953–1964. [Google Scholar]
  11. Zhou, Y.; Chen, Y.; Zhao, J.; Wu, Y.; Xu, J.; Li, J. What the role is vs. what plays the role: Semi-supervised event argument extraction via dual question answering. In Proceedings of the 35th AAAI conference on Artificial Intelligence, Online, 2–9 February 2021; pp. 14638–14646. [Google Scholar]
  12. Yang, S.; Feng, D.; Qiao, L.; Kan, Z.; Li, D. Exploring pre-trained language models for event extraction and generation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 5284–5294. [Google Scholar]
  13. Xu, N.; Xie, H.; Zhao, D. A novel joint framework for multiple Chinese events extraction. In Proceedings of the China National Conference on Chinese Computational Linguistics, Hainan, China, 30 October–1 November 2020; pp. 174–183. [Google Scholar]
  14. Huang, K.H.; Yang, M.; Peng, N. Biomedical event extraction with hierarchical knowledge graphs. arXiv 2020, arXiv:2009.09335. [Google Scholar]
  15. Zhang, X.; Zhu, Y.H.; OuYang, K.; Kong, L.W. Chinese Event Extraction Based on Role Separation. J. Shanxi Univ. 2022, 45, 936–946. [Google Scholar]
  16. Yang, H.J.; Jin, X.Y. A general model for entity relationship and event extraction. Comput. Eng. 2023, 49, 143–149. [Google Scholar]
  17. Zhu, M.; Mao, Y.C.; Cheng, Y.; Chen, C.J.; Wang, L.B. Event Extraction Method Based on Dual Attention Mechanism. Ruan Jian Xue Bao/J. Softw. 2023, 34, 3226–3240. [Google Scholar]
  18. Li, Q.; Li, J.; Sheng, J.; Cui, S.; Wu, J.; Hei, Y.; Peng, H.; Guo, S.; Wang, L.; Beheshti, A.; et al. A survey on deep learning event extraction: Approaches and applications. IEEE Trans. Neural Netw. Learn. Syst. 2022, 14, 1–21. [Google Scholar] [CrossRef]
  19. Li, F.; Peng, W.; Chen, Y.; Wang, Q.; Pan, L.; Lyu, Y.; Zhu, Y. Event extraction as multi-turn question answering. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, Online, 16–20 November 2020; pp. 829–838. [Google Scholar]
  20. Paolini, G.; Athiwaratkun, B.; Krone, J.; Ma, J.; Achille, A.; Anubhai, R.; Santos, C.N.d.; Xiang, B.; Soatto, S. Structured prediction as translation between augmented natural languages. In Proceedings of the Ninth International Conference on Learning Representations, Online, 3–7 May 2021. [Google Scholar]
  21. Hsu, I.; Huang, K.-H.; Boschee, E.; Miller, S.; Natarajan, P.; Chang, K.-W.; Peng, N. DEGREE: A data-efficient generation-based event extraction model. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Seattle, DC, USA, 10–15 July 2022; pp. 1890–1908. [Google Scholar]
  22. Van Nguyen, M.; Min, B.; Dernoncourt, F.; Nguyen, T. Joint extraction of entities, relations, and events via modeling inter-instance and inter-label dependencies. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Seattle, DC, USA, 10–15 July 2022; pp. 4363–4374. [Google Scholar]
  23. Liu, J.; Liang, C.; Xu, J.; Liu, H.; Zhao, Z. Document-level event argument extraction with a chain reasoning paradigm. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics, Toronto, ON, Canada, 9–14 July 2023; pp. 9570–9583. [Google Scholar]
  24. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  25. Yu, B.; Zhang, Z.; Sheng, J.; Liu, T.; Wang, Y.; Wang, Y.; Wang, B. Semi-open information extraction. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021; pp. 1661–1672. [Google Scholar]
  26. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  27. Su, J.; Lu, Y.; Pan, S.; Murtadha, A.; Wen, B.; Liu, Y. Roformer: Enhanced transformer with rotary position embedding. arXiv 2021, arXiv:2104.09864. [Google Scholar]
  28. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  29. Du, X.; Cardie, C. Document-level event role filler extraction using multi-granularity contextualized encoding. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 8010–8020. [Google Scholar]
  30. Zheng, S.; Wang, F.; Bao, H.; Hao, Y.; Zhou, P.; Xu, B. Joint extraction of entities and relations based on a novel tagging scheme. arXiv 2017, arXiv:1706.05075. [Google Scholar]
  31. Zeng, A.; Liu, X.; Du, Z.; Wang, Z.; Lai, H.; Ding, M.; Yang, Z.; Xu, Y.; Zheng, W.; Xia, X. Glm-130b: An open bilingual pre-trained model. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  32. Du, Z.; Qian, Y.; Liu, X.; Ding, M.; Qiu, J.; Yang, Z.; Tang, J. Glm: General language model pretraining with autoregressive blank infilling. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Dublin, Ireland, 22–27 May 2022; pp. 320–335. [Google Scholar]
Figure 1. Examples of three overlapping phenomena.
Figure 1. Examples of three overlapping phenomena.
Electronics 12 04688 g001
Figure 2. The overall framework of ROPEE.
Figure 2. The overall framework of ROPEE.
Electronics 12 04688 g002
Figure 3. Golden data of test case analysis on CasEE and ROPEE.
Figure 3. Golden data of test case analysis on CasEE and ROPEE.
Electronics 12 04688 g003
Table 1. Statistics of FewFC. Each column represents the size of each subset in terms of overlapped triggers, arguments, samples, and events.
Table 1. Statistics of FewFC. Each column represents the size of each subset in terms of overlapped triggers, arguments, samples, and events.
Trigger OverlapArgument OverlapSamplesEvent
Training13141541718510,277
Validation1682038991281
Testing1682098981332
All16501953898212,890
Table 2. Model comparative results of EE on four subtasks over FewFC. Highest scores are in bold.
Table 2. Model comparative results of EE on four subtasks over FewFC. Highest scores are in bold.
TI (%)TC (%)AI (%)AC (%)
PRF1PRF1PRF1PRF1
BERT-softmax89.879.084.080.261.869.874.662.868.272.560.265.8
BERT-CRF90.880.885.581.763.671.575.164.369.372.961.866.9
BERT-CRF-joint89.579.884.480.763.070.876.163.569.274.261.267.1
PLMEE83.785.884.775.674.575.174.367.370.672.565.568.8
MQAEE89.185.587.479.776.177.870.368.369.368.266.567.3
CasEE89.487.788.677.978.578.272.873.172.971.371.571.4
ROPEE88.888.288.574.782.878.669.578.673.867.676.972.0
Table 3. Comparative results of ROPEE and CasEE in two overlapping situations. Highest scores are in bold.
Table 3. Comparative results of ROPEE and CasEE in two overlapping situations. Highest scores are in bold.
TI (%)TC (%)AI (%)AC (%)
Trigger OverlapCasEE92.582.875.574.2
ROPEE92.082.476.475.4
Argument OverlapCasEE88.678.274.272.8
ROPEE89.278.774.973.4
Table 4. Comparative results on F1 using different role classifier strategies. Highest scores are in bold.
Table 4. Comparative results on F1 using different role classifier strategies. Highest scores are in bold.
TI (%)TC (%)AI (%)AC (%)
type_emb88.678.272.971.4
ROP-text_emb88.278.473.471.7
ROPEE88.578.673.872.0
Table 5. Comparative results on F1 using different training strategies. Highest scores are in bold.
Table 5. Comparative results on F1 using different training strategies. Highest scores are in bold.
TI (%)TC (%)AI (%)AC (%)
Role-1 88.177.671.769.4
Role-288.577.772.770.6
Role-388.478.173.571.7
Role-487.876.672.070.3
Role-588.277.272.470.7
ROPEE88.578.673.872.0
Table 6. Comparative results on F1 with and without trigger annotation.
Table 6. Comparative results on F1 with and without trigger annotation.
P (%)R (%)F1 (%)
ROPEEAI69.578.673.8
AC67.676.972.0
20 epoch w/o triggerAI71.472.672.0
AC70.071.070.5
50 epoch w/o triggerAI70.973.572.2
AC69.371.770.5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Q.; Yang, K.; Guo, X.; Wang, S.; Liao, J.; Zheng, J. Joint Overlapping Event Extraction Model via Role Pre-Judgment with Trigger and Context Embeddings. Electronics 2023, 12, 4688. https://doi.org/10.3390/electronics12224688

AMA Style

Chen Q, Yang K, Guo X, Wang S, Liao J, Zheng J. Joint Overlapping Event Extraction Model via Role Pre-Judgment with Trigger and Context Embeddings. Electronics. 2023; 12(22):4688. https://doi.org/10.3390/electronics12224688

Chicago/Turabian Style

Chen, Qian, Kehan Yang, Xin Guo, Suge Wang, Jian Liao, and Jianxing Zheng. 2023. "Joint Overlapping Event Extraction Model via Role Pre-Judgment with Trigger and Context Embeddings" Electronics 12, no. 22: 4688. https://doi.org/10.3390/electronics12224688

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop