Next Article in Journal
A Millimeter-Wave 4th-Harmonic Schottky Diode Mixer with Integrated Local Oscillator
Previous Article in Journal
Lactation Characteristics in Alpine and Nera di Verzasca Goats in Northern Italy: A Statistical Bayesian Approach
Previous Article in Special Issue
A Roadmap towards Breast Cancer Therapies Supported by Explainable Artificial Intelligence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

LSLSD: Fusion Long Short-Level Semantic Dependency of Chinese EMRs for Event Extraction

Department of Computer Science and Technology, Tongji University, Shanghai 200082, China
*
Author to whom correspondence should be addressed.
Submission received: 10 April 2021 / Revised: 30 July 2021 / Accepted: 4 August 2021 / Published: 5 August 2021
(This article belongs to the Special Issue Machine Learning in Medical Applications (Extended Version))

Abstract

:
Most existing medical event extraction methods have primarily adopted a simplex model based on either pattern matching or deep learning, which ignores the distribution characteristics of entities and events in the medical corpus. They have not categorized the granularity of event elements, leading to the poor generalization ability of the model. This paper proposes a diagnosis and treatment event extraction method in the Chinese language, fusing long short-level semantic dependency of the corpus, LSLSD, for solving these problems. LSLSD can effectively capture different levels of semantic information within and between event sentences in the electronic medical record (EMR) corpus. Moreover, the event arguments are divided into short word-level and long sentence-level, with the sequence annotation and pattern matching combined to realize multi-granularity argument recognition, as well as to improve the generalization ability of the model. Finally, this paper constructs a diagnosis and treatment event data set of Chinese EMRs by proposing a semi-automatic corpus labeling method, and an enormous number of experiment results show that LSLSD can improve the F1-value of event extraction task by 7.1% compared with the several strong baselines.

1. Introduction

Information extraction is a critical task in natural language processing, which aims at extracted structured information from unstructured or semi-structured texts. Medical research, such as disease source [1,2], disease prediction [3], and clinical decision [4,5], needs to extract structured information from underlying medical corpus. This kind of effective extraction and learning for knowledge has become a significant task. Traditionally, the medical knowledge graph (KG) takes entities as nodes, edges as relations to describe medical concepts. However, in addition to the involved medical concepts, the rich and closely related diagnosis and treatment events (DTE) in electronic medical records (EMRs) also contain dynamic knowledge and disease development logic, which cannot be captured by a fine-grained conceptual structure. Therefore, it is necessary to study the extraction method of DTE [6,7].
At present, automatic content extraction (ACE) [8] is the most famous event extraction conference. It divides event extraction (EE) into two subtasks: trigger word detection and event argument recognition. The event trigger word (ETW) is the word that can indicate the event type in the sentence, and the event argument (EA) is the participant that describes the event. For example, the sentence “On 12 August 2015, the patient underwent radical resection of rectal cancer in our hospital due to rectal cancer” is an operation event, “underwent” is an ETW, and the EAs include “12 August 2015”, “rectal cancer”, “radical resection of rectal cancer”.
Early research on medical DTE extraction mostly utilizes pattern matching [9,10]. Recent studies have shown that the sequence annotation methods based on the deep neural network have positive effects on the medical field [11,12]. These methods have incorporated the context of the entities in the form of distributed word representation and illustrated the importance of domain knowledge in clinical medical event extraction.
However, the basic-feature based EE methods merely adopt pattern matching or sequence annotation, which have limitations in fully mining the important but implied semantic dependency information contained in and between sentences. As shown in Table 1, both example 1 and example 2 are operation events. The ETWs and EAs contained in this type of event sentence have a certain regularity, such as that the ETWs are words of “underwent”. EAs often contain entity words such as disease and operation name, while example 3 of pathological examination usually takes a long sentence-level test result as one of the EAs, and it is mostly located after the operation event sentence.
In this paper, an elegant framework is proposed to recognize the complex ETWs and EAs in the rich DTEs in the history of present illness (HPI) texts of Chinese EMRs. The key idea of our work is to introduce different feature information to train the ETW classifier according to the distribution of entities and events at different corpus levels, and adopt different modules to realize the recognition of EAs with different granularity. Specifically, the contribution of this paper is three-fold:
  • Filling up the gap in the current representation framework and annotated corpus of HPI event in Chinese EMRs based on the combined source from the ACE task definition and Informatics for Integrating Biology and the Bedside (I2B2) [13], through adopting the reverse maximum matching algorithm performing semi-automatically labeling of the PME corpus of Chinese EMRs.
  • The proposed Chinese HPI event extraction method which fusing long short-level semantic dependency (LSLSD) of corpus not only highlights the semantic dependence between the components in sentences, but also emphasizes the semantic dependence between sentences. Meanwhile, it distinguishes the granularity of EAs at the short word-level and long sentence-level can better capture the semantic information of the corpus and enhance the generalization ability of the model.
  • A series of experiments on the annotated dataset show that the proposed model LSLSD has significant performance in trigger word detection and event argument recognition. It can effectively recognize EAs of different granularity at the same time.

2. Related Works

Event extraction [14] is an information extraction task that can be traced back to the 1980s. With the emergence of big data and the development of text mining and natural language processing, event extraction technology has been popularized. The technology of event extraction of the English context is mature [15], while that of the Chinese context, especially in clinical diagnosis and treatment events, is relatively limited. Zheng et al. [16] first proposed the language issues regarding Chinese event extraction, and divides event extraction into four sub-tasks of event recognition, event classification, argument recognition, and argument role classification. We follow ACE by dividing the event extraction into two sub-tasks of event trigger word detection and event argument recognition.

2.1. Event Trigger Word Detection

The approaches of the ETW detection task are mainly divided into two categories. One is the traditional pattern matching method based on syntactic analysis or the clustering method. Marco et al. [17] completed ETW detection of the open domain through rules of event extraction established by using syntactic analysis, Bui et al. [10] and Huang et al. [18] are completed ETW detection by building extraction rules according to the characteristics of biomedical text and using the method of joint constraint clustering, respectively. The other one is the current popular machine learning classification methods, which classify the ETWs after extracting them from sentences. For example, Xia et al. [19] detect ETW through multi-type features of lexical, dictionary, and syntactic according to the characteristics of biological literature, Yang et al. [20] and Zheng et al. [21] apply the remote supervision method to automatically label the financial texts from the knowledge base of Chinese financial event to realize the detection of financial ETWs. Jindal et al. [22] adopt the lexical features such as words and semantic relations in clinical terms to identify events. Wei et al. [23] construct a method based on sequence annotation and combine the static pre-training word vector represented by character level words with the dynamic context word representation based on the pre-training language model as the model input. However, most studies detect ETW by only using intra sentence information and ignore the information of inter sentences, while some scholars believe that the event distribution information at the document level can improve the accuracy of ETW detection of long-level semantic dependency. Xia et al. [24] construct event reasoning rules by using the document consistency feature to improve the classification results after recognizing and classifying ETW based on the semi-supervised model. Li et al. [25] utilize the combination semantics of ETWs and document consistency in Chinese training set to improve the performance of ETW detection, the former is used to infer unknown ETWs in the test set, and the latter is used to infer the situation that is difficult to deal with based on feature method.
To make the most of semantic information implied in the corpus, different from the previous models, we designed some extended features based on the basic features according to the characteristics of the data, including intra sentence features, inter sentence features, dependency syntax, and other features, to improve the accuracy of ETW detection model from the semantic perspective of entity information and relationship information.

2.2. Event Argument Recognition

Recently, the most popular methods of EA task have been based on sequence annotation; the model of bidirectional long short-term memory (BiLSTM) combined with conditional random field algorithm (CRF) is most frequently employed in sequence annotation task and has achieved excellent results. BiLSTM can capture useful context information from forward and backward of sentences, while CRF has the advantage of using sentence-level and adjacent labeled information when predicting current label. Xu et al. [26] construct an event model of clinical guidelines. The model extracts treatment events of Chinese clinical guidelines by through Word2Vec, LSTM and CRF technologies. Zeng et al. [27] propose a convolution BiLSTM neural network that combines LSTM and convolutional neural network (CNN) for Chinese event extraction, which can capture both sentence-level and local lexical information without any hand-craft features, in addition to the design of ETW location feature and ETW type feature and connect them with the original word vector. However, these models only extract treatment event contains short word-level arguments and do not divide EAs into long short-level granularity, and the arguments of long sentence-level are hardly extracted by a single sequence annotation model. Wei et al. [24] propose a multi-classification model based on self-attention, which fully utilizes the features of entity and entity type. However, the model only considers the granularity of the corpus when obtaining the representation vector. Jari et al. [28] focus on the edge detection, entity context features and the relation path feature between entities in the biomedical field. The hybrid method proposed by Xuan et al. [29] combines rule-based and feature-based classifier, which is effective in the long sentence-level EAs on biological EA recognition. We follow a classic method of BiLSTM-CRF [30] in the EA recognition task for short word-level arguments; the differences are that we employ the transformer [31] instead of the original encoder and we integrate extended semantic features and design a method of joint sequence annotation and pattern matching to recognize multi granularity EAs. For the Chinese EMR text purpose, we improve the generalization of the EA recognition model by dividing the EAs into long- and short-level granularity, and joint pattern matching and sequence annotation methods, so that the model can not only accurately identify the short word-level EAs in the corpus, but also effectively recognize the long sentence-level EAs widely existing in Chinese EMRs, to improve the performance of event argument recognition model.

3. Materials and Methods

The event extraction model in this paper is applied to the DTE extraction of the HPI texts in Chinese EMRs. In that sense, we complete the semi-automatic annotation of entity, ETW, and EA according to the annotation standard for entity and event form ACE and I2B2. Moreover, an event extraction model fusion long short-level semantic dependence of corpus has been proposed to capture the feature information of the corpus at different levels. Figure 1 illustrates our framework. To simplify the expression, the common nouns in the paper are abbreviated, as shown in Table 2. We will first introduce the semi-automatic annotation of the corpus.

3.1. Corpus Annotation

  • Data preprocessing: the HPI text of Chinese EMRs records the whole process of patients from illness to hospitalization and treatment, including symptoms, disease development, diagnosis results, detailed process of examination and treatment, etc. Considering the varied writing habits of doctors, the expression of the same disease, operation, or drug name may differ. There are abbreviations and variants which bring difficulties to the follow-up work. Therefore, the data need to be preprocessed. For example, the sentence segmentation for long sentences containing multiple events, deleting sentences unrelated to medical events, unifying medical vocabulary such as disease, operation, drug, standardizing disordered punctuation;
  • Defining event types and element arguments: I2B2 [13] held in 2012 defines a medical event as status, process, occurrence, and change, which are medically related. According to the definition of event extraction task in ACE and the definition of the medical event in I2B2, considering the content of the data set, the DTE of HPI are divided into six categories: admission, examination, test, treatment, operation, and diagnosis. The test events are subdivided into pathological test and immunohistochemistry, and the treatment events are subdivided into general treatment and chemotherapy. The event types and their event arguments are shown in Table 3;
  • Building a dictionary of candidate ETWs: the ETW is the core word that indicates the occurrence of a certain event in the sentence, generally verb or gerund. The words with high frequency in each type of events are selected in this paper as the candidate ETWs for each type of event. As shown in Table 3, the dictionary is expanded by using the synonyms of the candidate ETWs in the synonym forest;
  • Defining entity types: I2B2 and CCKS2017 (http://www.sigkg.cn/ccks2017/?page_id=51 (accessed on 18 January 2020)), respectively, divide medical named entities into three categories (medical problem, examination, treatment) and five categories (treatment, body part, symptom, examination, disease). Symptom entities are divided into body parts and symptom descriptions, and treatment entities into drug and operation in CCKS2018 (http://www.sigkg.cn/ccks2018/?page_id=16 (accessed on 18 January 2020)). Therefore, the entities in the data set are divided into three categories disease, symptom, and treatment. The symptom includes the body part and symptom description, and treatment includes drug, operation, and general treatment;
  • Corpus semi-automatic annotation: each HPI text contains 30–40 entities; manual annotation is time-consuming and labor-consuming. Therefore, this paper annotates corpus by a semi-automatic method based on a medical dictionary. Firstly, entity dictionary is collected from medical dictionaries (e.g., Dingxiangyuan (https://portal.dxy.cn/ (accessed on 21 September 2019)), Sogou Medical (https://pinyin.sogou.com/dict/cate/index/132 (accessed on 1 September 2019)) and 39 Health Net (http://www.39.net/ (accessed on 18 September 2019)), then a reverse maximum matching algorithm is employed to tag the texts automatically. Secondly, the ETWs, EAs, and ETW types in texts are manually annotated, referring to the definition of the event type. Finally, two doctors are invited to check and correct the above annotation results.

3.2. Event Trigger Word Detection Module

Event trigger word detection is a vital step in the event extraction task. Its result directly affects the accuracy of event extraction. In this paper, ETW detection is regarded as a multi-classification task. By analyzing the long short-level semantic dependence intra-event sentence and inter sentences in HPI texts, extended features are proposed to support the training of classifiers, and to obtain more accurate ETW detection results.
Firstly, the words existing in the sentence contained in the candidate ETW dictionary are condemned as the ETW of the sentence. If the words are not found, the ETW will be replaced by words with higher similarity in the sentence and the candidate ETW dictionary. Then, several types of features commonly utilized in ETW detection are selected as the basic features, as shown in Table 4 (1–4). For details, please refer to paper [32].
Based on the basic features suitable for general texts, extended features are designed for the entity distribution information in medical event sentences by analyzing the linguistic characteristic of Chinese HPI event sentences, as shown in Table 4 (5,6). Where feature f 5 and f 6 , respectively, refer to the distribution and number of entities in different types of event sentences. These two extended features of short-level semantic dependency are proposed because each type of event sentence does not cover all six types of entities, and the number of entities is not the same; these distribution regulars play an important role in the classification of ETWs.
In particular, there is a correlation between the events because HPI text describes the occurrence and evolution of the patient’s disease. For example, the admission event does not occur after the operation event, and the pathological examination event does not occur before the operation event. Therefore, inspired by [24,25], the extended feature of long-level semantic dependence refers to the consistency between event sentences in the HPI text is proposed named document consistency feature, as shown in Table 4 (7). Taking the operation event as an example, Figure 2 statistics the distribution of different event types before and after the operation event in the HPI texts. It can be seen that the probability of admission, diagnosis, and examination event occurring is higher before the operation event, but the probability of pathological test, chemotherapy, and general treatment event occurring is higher after the operation event.
Based on the above semantic features, for a given HPI text R , R = { S 1 , S 2 , , S n } , where S n = { W n 1 , W n 2 , , W n i } is the n th event sentence and W n i is the word in the sentence, where i express the sentence length (total number of words). The goal of ETW detection task is to obtain the classification P ( T | S n ) of ETW in the event sentence to determine the event type. Firstly, R is transformed into the corresponding word vector [ [ e n 1 W ] , [ e n 2 W ] , , [ e n i W ] ] through the pre-trained word embedding Word2Vec (https://github.com/tensorflow/tensorflow/tree/r1.1/tensorflow/examples/tutorials/word2vec (accessed on 22 April 2020)). Secondly, SVM is adopted as a classifier to classify ETWs. The one versus rest (OVR) method is adopted to realize the multi-classification task of ETW, K SVMs are trained for K ETWs categories, and the j th SVM is used to judge whether the ETWs belong to category j .
Specifically, SVM classifies the given x into category j by searching the maximum value of W j T x + b j . To mine the long short-level semantic dependence information in the corpus, we will concatenate the feature vector f m   ( m = 1 , 2 , , 7 ) of input HPI text R and the word vector [ [ e n 1 W ] , [ e n 2 W ] , , [ e n i W ] ] of R to obtain the semantic word vector S e :
S e = [ [ x 1 ] , [ x 2 ] , , [ x n i ] ] = [ [ e n 1 W ; f m ] , [ e n 2 W ; f m ] , , [ e n i W ; f m ] ]  
where ‘[;]’is the concatenation operation, S e d n i , x n i d x ^ ( d x ^ = d n i + d m ). Therefore, the dual form of the SVM optimization problem can be expressed as:
max : T = q = 1 n i α q 1 2 q , j = 1 n i α q α j y q y j K ( x q , x j )  
s . t . q = 1 n i α q y q = 0 ,   0 α q c
where α q , α j C , y q , y j are ETW types of x q and x j , respectively, K ( x q , x j ) is the kernel function, and c is the penalty coefficient.

3.3. Event Argument Extraction

Inspired by the research of [24], the EAs are divided into long sentence-level and short word-level according to the long short-level of the corpus, and the EAs are recognized by the method of joint sequence annotation and pattern matching. Here, the examination, pathological test, immunohistochemistry event result includes 20~50 words, defined as long sentence-level EA, and other EAs are short word-level EA.
In addition to the intra sentence and inter sentence semantic features described in Section 3.2, there are other semantic features in the event sentence of HPI, such as the location feature and type feature of ETWs and the dependency syntactic feature of event sentences shown in Table 5. The location feature refers to the distance between EAs and ETW in event sentence can provide syntactic information about the event for EA recognition because the distribution of EAs usually surrounds ETW. The ETW position is encoded by the transformer, and the even and odd position is encoded by P E ( p o s , 2 i ) = sin ( p o s / 10 , 000 2 i / d ) and P E ( p o s , 2 i + 1 ) = cos ( p o s / 10 , 000 2 i / d ) , respectively, where p o s is the ETW position in the sequence, d and i are vector dimension, and i th item in the sequence, respectively.
Meanwhile, different event types often correspond to different kinds of EAs, which are closely related, while ETW type implies the type of event sentence. Moreover, under the writing standard of HPI text, the same type of event sentences usually follows a similar syntactic structure, such as the treatment event sentence “On May 14, 2015, the patient was given acid making, anti-inflammatory, and antiemetic nursing.” and “The patient was given treatment for symptomatic after the operation, such as protecting liver and stomach and improving immunity.”. The former ETW is “nursing”, EAs are “acid making”, “anti-inflammatory” and “antiemetic”, the latter ETW is “treatment for symptomatic”, EAs are “protecting liver” and “protecting stomach” and “improving immunity”. Although the ETWs of the two event sentences are not the same, the EAs in each sentence are juxtaposed, and the EA and ETW are modifier-head constructions. Thus, the syntactic components such as “subject predicate object” and “attribute adverbial complex” in the event sentence can be identified by dependency parsing. The syntactic relations of dependency parsing, and their corresponding label are shown in Table 6; the specific types of relations can refer to the language technology platform LTP.
(1) For EAs of short word-level, recognized by the sequence annotation method fusion semantic feature vectors of corpus, the model framework is given in Figure 3. Firstly, the vector representation [ [ e n 1 B ] , [ e n 2 B ] , , [ e n i B ] ] of all words in each event sentence from HPI text R is obtained by BERT [33]. Secondly, the word vector is fused with semantic feature vectors f m ( m = 5 , 8 , 9 , 10 ); the fusion process is as follows:
F 1 ( e n 1 B , f m ) = e n i B = [ [ e n 1 B ; f m ] , [ e n 2 B ; f m ] ,   ,   [ e n i B ; f m ] ]
Then, encoding the fused vectors by a bi-directional LSTM network with a hidden layer size of d n ˜ can learn how to judge the key fusion vectors to obtain the semantic information of short word-level, where d n ˜ = 1 / 2 d n i . In the t th time step, the hidden layer state of BiLSTM output is h n i t with the input e ^ n i t ; this process is as following:
[ h n 1 L ] ,   [ h n 2 L ] , ,   [ h n i L ] = B i L S T M ( [ e ^ n 1 L ] , [ e ^ n 2 L ] , , [ e ^ n i L ] )
The dropout layer is adopted to make some bi-directional long short-term memory units randomly inactivated for avoiding over-fitting of training results, and the semantic information of short word-level is enhanced by concatenating the vectors fused semantic features.
F 2 ( h n i L , r n i L , e ^ n i L ) = f ( W L r L ( h n i L e ^ n i L ) + b L )
where W L d n i × d x ^ is the weight matrix, b L is bias, and f ( ) is activation function, r L = [ r n 1 L , , r n i L ] , r n i L Bernoulli ( p ) . Here, Bernoulli is utilized to randomly generate a vector of 0 or 1.
Finally, the output h ^ n i L of dropout is entered into CRF and the corresponding label sequence is y a = ( y n 1 a , , y n i a ) . Then, all parameters of CRF can be estimated by maximizing P ( y a | h ^ L ) for the given HPI text R :
P ( y a | h ^ L ) = 1 Z ( h ^ n i L ) exp   ( g = 1 n i v = 1 n i   ( λ g E y g a , g + μ v T ˜ y v 1 a , y v a ) )
Z ( h ^ n i L ) is the normalization factor, E y g a , g is the probability of label y g a corresponds to h ^ g L , T ˜ y v 1 a , y v a is the probability of label y v 1 a corresponds to h ^ v 1 L under the premise of label y v a corresponds to h ^ v L , λ g and μ v are hyperparameters. Therefore, the CRF is trained by calculating the maximum log-likelihood function:
max : A = ( S , y a ) R l o g P ( y a | h ^ L )
To sum up, we train the event extraction model LSLSD in Chinese DTE by minimizing the following loss function:
( Θ ) = ( λ 1 T + λ 2 A )  
where λ 1 and λ 2 are weights, Θ is the sum of the parameters of the model LSLSD.
(2) For long sentence-level EAs, being recognized by the pattern matching rules that are given in Table 7 is not only resulted from the low detection accuracy of machine learning methods as the EAs contain a generous number of words, but also because the sentence structure of examination and test event corresponding to long sentence-level EA is relatively regular and has distinct syntactic characteristics. For example, pathological test (PT) event “On 3 June 2015, postoperative pathology showed: colon cancer, infiltrating to the upper serosa, perirectal…adenocarcinoma metastasis”, where the ETW is “pathology” and the EAs are time argument “3 June 2015”, disease argument “colon cancer” and test result argument “infiltrating to the upper serosa…adenocarcinoma metastasis”. From the dependency syntactic analysis of the event sentence, Figure 4a, time argument and “showed” have adverbial structure, while “showed” has the subject–predicate structure with ETW “pathology”; the disease argument has the verb–object structure with “showed”; the long sentence-level test result argument contains all words in the recursive juxtaposition structure with disease argument. The logical expression of the rule is shown in Table 7-E1. The dependency syntactic analysis of immunohistochemistry (IHC) event “On 12 April 2016, underwent immunohistochemistry, reported as KI50%, ER60%, …, SYN+, CK10%” is shown in Figure 4b, where the ETW is “immunohistochemistry” and the EAs are time argument “12 April 2016” and test result argument “KI50%, ER60%, …, SYN+, CK10%”. The logical expression of the rule for such sentence pattern is shown in Table 7-IHC1. “CT”, “4 September 2016”, “esophageal cancer” and “showing local thickening of mid-esophageal, uneven thickening of the lower esophagus” in CT examination event “On 4 September 2016, take CT showed: esophageal cancer, local thickening of middle esophageal, uneven thickening of lower esophagus” are the ETW, time argument, disease argument, and examination result argument, respectively, and its dependency syntactic analysis and logical expression of rules are shown in Figure 4c and Table 7-E1, respectively.

4. Experiments

4.1. Dataset

Annotating the data from 1000 HPI texts of Chinese EMRs in a top-three hospital in Shanghai by the semi-automatic annotation method proposed in Section 3.1, and the data set of Chinese medical event (CME_1000) is obtained, which contains 7035 event sentences. The number of ETWs and EAs in different event sentences is shown in Table 8; “#” represents the number. Here, 70%, 20%, and 10% of the CME_1000 are randomly adopted as the training set, test set, and validation set, respectively, and there are no duplicate items in these three sets.

4.2. Details

There are several comparative experiments designed for the subtask of event trigger word detection and event argument recognition of event extraction; the experiment results are evaluated by the index value of precision (P), recall (R), and F1-value.
In the event trigger word detection module, the software package LibSVM [34] is adopted to train and test the SVM; select RBF kernel for K ( x q , x j ) in the formula (2); the constant value coefo of kernel function is 0; the penalty coefficient c is 1.
In the event argument recognition module, the neural network framework Tensorflow 2.0 (https://www.tensorflow.org/install (accessed on 13 October 2019)) is employed for model training and the pre-training language model adopts BERT_base. Adam optimizer is used in the BiLSTM layer, and the batch size, epochs, learning rate, and dropout rate are set to 16, 20, 2 × 10−5, and 0.5, respectively.

4.3. Results

In the aspect of ETW detection, aiming at proving the positive effect of the proposed extended features fusion of long short-level semantic dependency on the detection results, the model performance of adding extended features of short-level semantic dependency is compared based on the baseline model without fusion of extended features (only feature f 1 ~ f 4 are included). The performance of the ETW detection module under different feature combinations on CME_1000 is shown in Table 9. From Table 9, the experiment results of the baseline are not ideal because the entity information in the medical text is not fully utilized. After adding the entity type feature f 5 (T) of intra sentence, the classification results of all kinds of events are improved because different entity types only appear in specific event types. For example, the P and F1 value of the operation event increased by 5.3% and 3.61%. This is because the ETW of operation event often appears in chemotherapy, test, and examination event, and the entities of operation type only appear in admission, diagnosis, and operation event, ensuring the entity type feature T of intra sentence can improve the ETW detection results of operation event. Moreover, after adding the number feature f 6 (N) of different entities of inter sentences, the average F1 value for all events increased from 81.07% to 82.45%. For example, the F1 value of the examination event is increased by 3.44%. This is because the feature N can effectively distinguish test and examination events with similar entity type distribution.
In the aspect of EA extraction, our model LSLSD is compared with word-based C-BiLSTM and char-based C-BiLSTM proposed by Zeng et al. [27] to prove the better performance of LSLSD in EA extraction. As shown in Table 10, the short word-level arguments and long sentence arguments of the corpus are better captured by LSLSD, which joins sequence annotation and pattern matching. In the EA recognition task, the F1 value of LSLSD is 6.5% and 5.1% higher than word-based C-BiLSTM and char-based C-BiLSTM respectively. The F1 value of LSLSD is 11.2% and 8.2% higher than word-based C-BiLSTM and char-based C-BiLSTM, respectively, in the EA classification task.

4.4. Discussion

In terms of ETW detection, the low accuracy of ETW detection is directly reduced by the ETWs that infrequently appear in the candidate ETW dictionary that exists in HPI texts. The accuracy of ETW module is improved by setting a two-stage experiment; the first stage determines whether each candidate ETW, verb, and gerund are ETW through direct binary classification by using features of word, part of speech, and the number of different entities type. The second stage determines the type by multi-classification for the ETWs selected in the first stage. As shown in Table 11, the F1 value of the two-stage ETW detection has been significantly improved under the same feature combination. For example, compared with the result in Table 10, the F1 value of the chemotherapy event increased from 89.91% to 91.48% through the two-stage method under the feature combination of “Baseline + T”.
In addition, the event sentence with the accuracy of TWE classification is less than 40% by fusing extended features of short-level semantic dependence named untrusted event sentence; otherwise, it is a trusted event sentence. The untrusted event sentences are classified a second time by adding long-level semantic dependence features f 7 and other extended features of enter sentences to improve the accuracy of ETW detection. From the ETW detection results under different features combinations in Table 12, the result of the “P1 + R1 + B + N + PN + RN” combination is the best. Compared with the two-stage ETW detection method, the R and F1 value of experiment with document consistency features are improved from 79.03% to 80.61% and from 82.23% to 83.57%, respectively, which indicate that document consistency features can supplement the information difficult to extract on the long sentence-level and improve the performance of the model. The trend of F1 value change of different events under different feature combinations are shown in Figure 5, in which the entity type feature T of intra sentence significantly improves the ETW detection effect on chemotherapy, treatment, operation, and examination events, and the number feature N of different entities of inter sentences significantly improves the ETW detection effect on the pathological test, immunohistochemistry, operation, and examination events. The two-stage experiment had a great influence on treatment, operation, examination, and chemotherapy events, while the document consistency features S improved the results of ETW detection of all event types.
In terms of EA extract, the influence of different word vector pre-training models of Word2Vec, ELMO, and BERT on the recognition of ETW and EA are compared in this paper; as shown in Table 13, the effect of BERT is better than that of Word2Vec and ELMO, which is because Word2Vec does not consider the influence of context information on the word vector when constructing the text vector so that the word vector generated by the same word is invariant. ELMO employs stacked BiLSTM to make the generated word vectors contain more context information of the vocabulary, so the ELMO is better than Word2Vec. The model effect of using BERT is better than ELMO and Word2Vec, due to the generated text sequence of BERT having deeper context information and lexical-semantic by feature extractor of transformer. The experiment results of removing the CRF model are obtained to highlight the positive effect of CRF on improving the accuracy of label prediction by adding constraint rules, and the results show that the effect of using BERT-BiLSTM along is far less than that of adding the CRF model.
In addition, the ablation experiment is designed to better study the influence of the four extra input vectors of BiLSTM on the EA recognition task. This experiment utilizes the sequence annotation model of BERT-BiLSTM-CRF for all event types, the experiment results are shown in Table 14, where D represents the distance vector between the word and the ETW, and C and Y represent the type vector of the ETW and the dependency syntactic information vector, respectively.
From Table 14, the addition of four types of vectors improves the performance of the model in varying degrees, which proves the effectiveness of the proposed feature vectors. It is noted that the experimental effect is higher than that of BERT-BiLSTM-CRF when the dependency syntactic information vector Y is added into the combination of the distance vector D, type vector C of ETW and entity type vector T of intra sentence, but it is slightly lower than before. The same situation occurs again for the vector combination of DC and DCY, which illustrates that the addition of dependency syntactic information vector will harm the combination of other features that may be caused by the characteristics of syntactic structure is not obvious due to the different writing styles of different doctors, although the medical text has a certain degree of standardization. Compared with the BERT-BiLSTM-CRF, the F1 values of the best combination DCT among all vector combinations on EA recognition and EA classification tasks are improved by 5.9% and 4.5%, respectively.

5. Conclusions

A novel model for event extraction on Chinese history of present illness texts by fusing its long short-level semantic dependency is proposed in this paper, and the experiment results demonstrated that the proposed extended features with implicit information of intra-sentence or inter-sentences help improve the performance of LSLSD. Meanwhile, LSLSD has excellent generalization performance, due to the fact that the event arguments recognition module divides event arguments into the long sentence-level and short word-level, and it recognizes event arguments by the method of joint sequence annotation and pattern matching.
We also noticed in the experiment results analysis that, overall, the dependence syntactic feature proposed in this paper has negative effects on some event arguments recognition. The reason for this effect is that the HPI text data set is small; on the other hand, the writing styles of different clinicians are different. Further improvement might be made by expanding the scale of Chinese HPI texts, to explore more profound semantic features. We will study advanced deep learning algorithms to improve the accuracy of event argument recognition.

Author Contributions

Methodology, P.Z., C.W.; software, P.Z., C.W.; formal analysis, P.Z., Y.F.; investigation, P.Z., C.W.; data curation, P.Z., C.W.; writing—original draft preparation, P.Z., C.W.; writing—review and editing, P.Z., Y.F.; visualization, P.Z., C.W.; supervision, Y.F.; project administration, Y.F.; funding acquisition, Y.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China (No. 2019YFB2101600).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The electronic medical record data obtained from the hospital has not been made available due to privacy concerns.

Acknowledgments

We would like to thank two doctors from a First-class Hospital at Grade III in Shanghai, who manually checked and corrected the data set that is labeled by our semi-automatic annotation method for the experiments in this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Guo, S.; Jin, Z.; Gotz, D.; Du, F.; Zha, H.; Cao, N. Visual progression analysis of event sequence data. IEEE Trans. Vis. Comput. Graph. 2018, 25, 417–426. [Google Scholar] [CrossRef] [PubMed]
  2. Kwon, B.C.; Anand, V.; Severson, K.A.; Ghosh, S.; Sun, Z.; Forhnert, B.I.; Lundgren, M.; Ng, K. DPVis: Visual exploration of disease progression pathways. arXiv 2019, arXiv:1904.11652. [Google Scholar]
  3. Jin, Z.; Cui, S.; Guo, S.; Gotz, D.; Sun, J.; Cao, N. Carepre: An intelligent clinical decision assistance system. ACM Trans. Comput. Healthc. 2020, 1, 1–20. [Google Scholar] [CrossRef] [Green Version]
  4. Yazhini, K.; Loganathan, D. A state of art approaches on deep learning models in healthcare: An application perspective. In Proceedings of the 2019 3rd International Conference on Trends in Electronics and Informatics (ICOEI), IEEE, Tirunelveli, India, 23–25 April 2019; pp. 195–200. [Google Scholar]
  5. Harerimana, G.; Kim, J.W.; Yoo, H.; Jang, B. Deep learning for electronic health records analytics. IEEE Access 2019, 7, 101245–101259. [Google Scholar] [CrossRef]
  6. Yadav, S.; Ramteke, P.; Ekbal, A.; Saha, S.; Bhattacharyya, P. Exploring disorder-aware attention for clinical event extraction. ACM Trans. Multimed. Comput. Commun. Appl. 2020, 16, 1–21. [Google Scholar] [CrossRef]
  7. Fócil-Arias, C.; Sidorov, G.; Gelbukh, A. Medical events extraction to analyze clinical records with conditional random fields. J. Intell. Fuzzy Syst. 2019, 36, 4633–4643. [Google Scholar] [CrossRef]
  8. Doddington, G.R.; Mitchell, A.; Przybocki, M.A.; Ramshaw, L.A.; Strassel, S.M.; Weischedel, R.M. The automatic content extraction (ace) program-tasks, data, and evaluation. In Proceedings of the 4th International Conference on Language Resources and Evaluation, Lisbon, Portugal, 26–28 May 2004; Volume 2, pp. 837–840. [Google Scholar]
  9. Kovačević, A.; Dehghan, A.; Filannino, M.; Keane, J.A.; Nenadic, G. Combining rules and machine learning for extraction of temporal expressions and events from clinical narratives. J. Am. Med. Inform. Assoc. 2013, 20, 859–866. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Bui, Q.C.; Campos, D.; van Mulligen, E.; Kors, J. A fast rule-based approach for biomedical event extraction. In Proceedings of the BioNLP Shared Task 2013 Workshop, Sofia, Bulgaria, 9 August 2013; pp. 104–108. [Google Scholar]
  11. Viani, N.; Miller, T.A.; Napolitano, C.; Priori, S.G.; Savova, G.K.; Bellazzi, R.; Sacchi, L. Supervised methods to extract clinical events from cardiology reports in Italian. J. Biomed. Inform. 2019, 95, 103219. [Google Scholar] [CrossRef]
  12. Zhou, X.; Xiong, H.; Zeng, S.; Fu, X.; Wu, J. An approach for medical event detection in Chinese clinical notes of electronic health records. BMC Med. Inform. Decis. Mak. 2019, 19, 31–37. [Google Scholar] [CrossRef]
  13. Uzuner, Ö.; Luo, Y.; Szolovits, P. Evaluating the state-of-the-art in automatic deidentification. J. Am. Med. Inform. Assoc. 2007, 14, 550–563. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Hogenboom, F.; Frasincar, F.; Kaymak, U.; De Jong, F.; Caron, E. A survey of event extraction methods from text for decision support systems. Decis. Support. Syst. 2016, 85, 12–22. [Google Scholar] [CrossRef]
  15. Devlin, J.; Zbib, R.; Huang, Z.; Lamar, T.; Schwartz, R.; Makhoul, J. Fast and robust neural network joint models for statistical machine translation. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, Baltimore, MD, USA, 23–25 June 2014; Volume 1, pp. 1370–1380. [Google Scholar]
  16. Chen, Z.; Ji, H. Language specific issue and feature exploration in Chinese event extraction. In Proceedings of the Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics, Boulder, CO, USA, 31 May–5 June 2009; pp. 209–212. [Google Scholar]
  17. Valenzuela-Escárcega, M.A.; Hahn-Powell, G.; Surdeanu, M.; Hicks, T. A Domain-independent rule-based framework for event extraction. In Proceedings of the ACL-IJCNLP 2015 System Demonstrations, Beijing, China, 26–31 July 2015; pp. 127–132. [Google Scholar]
  18. Huang, L.; Cassidy, T.; Feng, X.; Ji, H.; Voss, C.; Han, J.; Sil, A. Liberal event extraction and event schema induction. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 7–12 August 2016; Volume 1, pp. 258–268. [Google Scholar]
  19. Xia, J.; Fang, A.C.; Zhang, X. A novel feature selection strategy for enhanced biomedical event extraction using the Turku system. BioMed Res. Int. 2014, 2014, 205239. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Yang, H.; Chen, Y.; Liu, K.; Xiao, Y.; Zhao, J. DCFEE: A document-level Chinese financial event extraction system based on automatically labeled training data. In Proceedings of the ACL 2018, System Demonstrations, Melbourne, Australia, 15–20 July 2018; pp. 50–55. [Google Scholar]
  21. Zheng, S.; Cao, W.; Xu, W.; Bian, J. Doc2EDAG: An end-to-end document-level framework for chinese financial event extraction. arXiv 2019, arXiv:1904.07535. [Google Scholar]
  22. Jindal, P.; Roth, D. Extraction of events and temporal expressions from clinical narratives. J. Biomed. Inform. 2013, 46, S13–S19. [Google Scholar] [CrossRef] [Green Version]
  23. Wei, Y.; Liu, M.; Hu, H. Biomedical event extraction based on deep contextual word representation and self-attention. Comput. Eng. Sci. 2020, 42, 1670–1679. (In Chinese) [Google Scholar]
  24. Xu, X.; Li, P.F.; Zheng, X.; Zhu, Q. Event inference for semi-supervised Chinese event extraction. J. Shandong Univ. Nat. Sci. 2014, 12, 3. [Google Scholar]
  25. Li, P.; Zhou, G.; Zhu, Q.; Hou, L. Employing compositional semantics and discourse consistency in Chinese event extraction. In Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, Jeju, Korea, 12–14 July 2012; pp. 1006–1016. [Google Scholar]
  26. Yu, H.; Xu, C.; Liu, Y.; Fu, Y.; Gao, D. BiLSTM and CRF-based extraction of therapeutic events from Chinese clinical guidelines. Chin. J. Med. Libr. Inf. Sci. 2020, 29, 9–14. (In Chinese) [Google Scholar]
  27. Zeng, Y.; Yang, H.; Feng, Y.; Wang, Z.; Zhao, D. A convolution BiLSTM neural network model for Chinese event extraction. In Natural Language Understanding and Intelligent Applications; Springer: Kunming, China, 2–6 December 2016; pp. 275–287. [Google Scholar]
  28. Björne, J.; Salakoski, T. Generalizing biomedical event extraction. In Proceedings of the BioNLP Shared Task 2011 Workshop, Portland, OR, USA, 24 June 2011; pp. 183–191. [Google Scholar]
  29. Pham, X.Q.; Le Minh, Q.; Ho, B.Q. A hybrid approach for biomedical event extraction. In Proceedings of the BioNLP Shared Task 2013 Workshop, Sofia, Bulgaria, 9 August 2013; pp. 121–124. [Google Scholar]
  30. Huang, Z.; Xu, W.; Yu, K. Bidirectional lstm-crf models for sequence tagging. arXiv 2015, arXiv:1508.01991. [Google Scholar]
  31. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  32. Wang, C.; Zhai, P.; Fang, Y. Chinese medical event detection based on feature extension and document consistency. In Proceedings of the 5th International Conference on Automation, Control and Robotics Engineering (CACRE), IEEE, Dalian, China, 18–20 September 2020; pp. 753–758. [Google Scholar]
  33. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  34. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
Figure 1. The framework of LSLSD. This model is composed of two modules: trigger word detection (A) and event argument recognition (B) (the pink and blue background color, respectively). f 1 ~ f 4 are the basic features commonly used in trigger word detection task; f 5 ~ f 10 are the extended features proposed in this paper. By fusing these semantic features, the potential association information within and between the event sentences in HPI texts can be fully mined.
Figure 1. The framework of LSLSD. This model is composed of two modules: trigger word detection (A) and event argument recognition (B) (the pink and blue background color, respectively). f 1 ~ f 4 are the basic features commonly used in trigger word detection task; f 5 ~ f 10 are the extended features proposed in this paper. By fusing these semantic features, the potential association information within and between the event sentences in HPI texts can be fully mined.
Applsci 11 07237 g001
Figure 2. Distribution of event types before and after operation events.
Figure 2. Distribution of event types before and after operation events.
Applsci 11 07237 g002
Figure 3. Recognition model for short word-level by fusing semantic feature vectors.
Figure 3. Recognition model for short word-level by fusing semantic feature vectors.
Applsci 11 07237 g003
Figure 4. The dependency syntactic analysis of event sentences.
Figure 4. The dependency syntactic analysis of event sentences.
Applsci 11 07237 g004
Figure 5. F1-value variation tendency under different feature combinations.
Figure 5. F1-value variation tendency under different feature combinations.
Applsci 11 07237 g005
Table 1. Examples of event sentences from Chinese present medical history.
Table 1. Examples of event sentences from Chinese present medical history.
Example #1
ES 1: On 12 August 2015, the patient underwent radical resection of rectal cancer in our hospital due to “rectal cancer”.
ET 2: operation event ETW 3: underwent
EA 4: 12 August 2015, rectal cancer, radical resection of rectal cancer
Example #2
ES: On 1 January 2014, the patient underwent “radical resection of gastric stump cancer” under general anesthesia in our hospital due to “gastric stump cancer”.
ET: operation event ETW: underwent under general anesthesia
EA: 1 January 2014, gastric stump cancer, radical resection of gastric stump cancer
Example #3
ES: On 25 June 2015, the postoperative pathology showed: “moderately differentiated adenocarcinoma of sigmoid colon, infiltrating into the upper mucosa of intestinal wall, without nerve bundle invasion, one intermediate lymph node metastasis (1/7), central lymph node (0/5), para intestinal lymph node (0/2)”.
ET: pathological test event ETW: pathology
EA: 25 June 2015, moderately differentiated adenocarcinoma of sigmoid colon…para intestinal lymph node (0/2)
1 ES, 2 ET, 3 ETW, and 4 EA stand for event sentence, event type, trigger, and argument.
Table 2. Abbreviations of nouns are commonly used in the paper.
Table 2. Abbreviations of nouns are commonly used in the paper.
NounsAbbreviations
Diagnosis and Treatment EventDTE
Event ExtractionEE
Event Trigger WordETW
Event ArgumentEA
History of Present IllnessHPI
Table 3. Examples of representation framework and trigger works for HPI events.
Table 3. Examples of representation framework and trigger works for HPI events.
Event TypesSub-TypesEAsExample of ETWs
AdmissionAdmissionTime, Synonym,
Operation name, Disease
admission, see a doctor,
hospitalization, receive and cure
TestPathological testTime, Disease,
Result of pathological test
pathology, biopsy, slicing,
pathological examination
ImmunohistochemistryTime, Result of immunohistochemistryimmunological staining,
staining test
ExaminationExaminationTime, Disease,
Result of examination
gastroscopy, blood routine,
CT, colonoscopy
TreatmentGeneral treatmentTime, treatment methodadjuvant therapy, healing,
nursing
ChemotherapyTime, Drugchemotherapy, targeted therapy, radiation
OperationOperationTime, Synonym,
Operation name, Disease
underwent, surgery,
general anesthesia
DiagnosisDiagnosisTime, Synonym,
Operation name, Disease
diagnosis, judgment,
determine, results
Table 4. Semantic features faced to event trigger word detection.
Table 4. Semantic features faced to event trigger word detection.
Feature TypesNumberSymbolsNamesDescribes
Basic feature1 f 1 Trigger wordWord vector by LTP 1 word segmentation tool
2 f 2 Part-of-speechPart-of-speech tagged by LTP
3 f 3 contextWord and part-of-speech in window of length 3
4 f 4 dictionaryEvent types corresponding to candidate ETW dictionary
Extended featureIntra sentence5 f 5 Entity typesDistribution of entities in different types of event sentences
6 f 6 #2 different entity typesNumber of entities in different types of event sentences
Inter sentences7 f 7 Document consistencyDistribution of entity types before and after event sentences
1 LTP is a language technology platform constructed by Harbin Institute of Technology (http://ltp.ai/ (accessed on 6 March 2020)); 2# is the number.
Table 5. Semantic features faced to argument recognition.
Table 5. Semantic features faced to argument recognition.
Feature TypesNumberSymbolsNamesDescribes
Extended
features
8 f 8 Location of ETWDistance between EAs and ETW
9 f 9 Type of ETWTypes of ETW obtained by module A
10 f 10 Dependency syntacticDependency syntactic in event sentences
Table 6. Relations under dependency syntactic.
Table 6. Relations under dependency syntactic.
Relation TypesLabelsRelation TypesLabels
Subject–predicate relationSBVVerb–complement constructionCMP
Verb–object relationVOBJuxtapositionCOO
Indirect object relationIOBPreposition-object relationPOB
Preposed-objectFOBLeft additional relationLAD
Concurrent remarkDBLRight additional relationRAD
Modifier–head constructionATTIndependent structureIS
Mesomorphic constructionADVPunctuationW
Table 7. Relations under dependency syntactic.
Table 7. Relations under dependency syntactic.
Event TypesLogical Expression
E1 τ A D V   V O B T V O B V O B   A D V D       w     RS 1, T V O B V O B   C O O RE 2
E2T w RS S B V   C O O C O O RE V O B D
E3 τ A D V A D V T A D V   V O B D       w     RS A D V   A T T A T T   A D V RE
E4 τ A D V   V O B T w RS w C O O C O O RE
PT1 τ A D V S B V T S B V V O B D C O O RS C O O C O O RE
PT2T S B V   V O B D       w     RS C O O S B V D S B V V O B RE
PT3 τ A T T A T T T A T T S B V         w     RS C O O C O O RE
PT4T A T T   w RS A T T A T T   C O O C O O RE C M P P O B D
PT5T A T T   CMP   D w RS A T T   C O O C O O RE
IHC1 τ A D V   V O B T V O B     C O O   V O B RS C O O C O O RE
IHC2T S B V   V O B RS C O O   w   C O O   w   RE
IHC3T S B V A D V τ , T S B V V O B RS C O O C O O RE
1 RS, 2 RE are the beginning and end of the result argument, respectively.
Table 8. Statistics of CME_1000 for event extraction in Chinese HPI text.
Table 8. Statistics of CME_1000 for event extraction in Chinese HPI text.
Dataset# HPI Text# Entity# Event Sentence# ETW# EA
Training70016,0714822482214,296
Validation10021947287281681
Test2003528148514853607
Total100021,7937035703519,584
Table 9. Results of trigger detection under different feature combinations.
Table 9. Results of trigger detection under different feature combinations.
Event TypesChemotherapyExamination
Methods PRF1PRF1
Baseline94.7182.0687.9387.1460.3971.34
Baseline + T95.2985.1189.9187.4563.2373.39
Baseline + T + N95.3186.0690.4489.4767.3276.83
Event TypesOperationGeneral Treatment
Methods PRF1PRF1
Baseline68.9686.2576.6486.9876.9581.66
Baseline + T74.2687.2880.2587.2281.4884.25
Baseline + T + N75.8987.6281.3387.2481.5784.31
Event TypesAdmissionImmunohistochemistry
Methods PRF1PRF1
Baseline78.3473.4475.8169.9176.5173.06
Baseline + T81.2275.4778.2470.2781.2975.38
Baseline + T + N82.1676.5079.2375.4682.6078.87
Event TypesPathological TestDiagnosis
Methods PRF1PRF1
Baseline87.8477.5182.3586.9181.7784.26
Baseline + T88.1377.2882.3589.0980.9084.80
Baseline + T + N91.0676.8183.3391.2680.0585.29
Table 10. Results of event argument extraction under different methods.
Table 10. Results of event argument extraction under different methods.
Event TypesEA RecognitionEA Classification
Methods PRPRPR
Word-based C-BiLSTM 59.852.759.852.759.852.7
Char-based C-BiLSTM 61.258.561.258.561.258.5
LSLSD (our)66.367.566.367.566.367.5
Table 11. Results of trigger detection by two-stage method.
Table 11. Results of trigger detection by two-stage method.
Event TypesChemotherapyExamination
Methods PRF1PRF1
Two-stageBaseline + T95.0688.1691.4888.6169.3177.78
Baseline + T + N95.1288.3291.5989.8770.2978.88
Event TypesOperationGeneral treatment
Methods PRF1PRF1
Two-stageBaseline + T77.8785.9181.6987.8783.5385.65
Baseline + T + N80.7587.9784.2187.8684.3185.91
Event TypesAdmissionImmunohistochemistry
Methods PRF1PRF1
Two-stageBaseline + T84.3375.6679.7676.0583.6579.67
Baseline + T + N85.2076.6280.6876.1281.8778.89
Event TypesPathological TestDiagnosis
Methods PRF1PRF1
Two-stageBaseline + T90.7175.6782.5191.6480.2285.55
Baseline + T + N91.4576.9183.5591.5679.0684.85
Table 12. Results of trigger detection based on long-level semantic dependency feature.
Table 12. Results of trigger detection based on long-level semantic dependency feature.
Event TypesALL-TOTAL
Methods PRF1
B 1 + P1 2 + R1 386.3179.4282.72
B + P1 + R1 + T + PT 4 + RT 586.5480.1383.21
B + P1 + R1 + T + N + PTN 6 + RTN 785.8479.3682.47
B + P1 + R1 + N + PN 8 + RN 986.9780.6183.57
1 B represents the basic features. 2 P1, 3 R1, 4 PT, 5 RT, 6 PTN, 7 RTN, 8 PN and 9 RN are the document consistency feature, where P1 and R1 represent the event type of the previous and latter sentence respectively; PT and RT, respectively, represent the feature (T) of entity type in the previous and latter sentence; PN and RN are the number feature (N) of different entity types in the previous and latter sentence; and PTN and RTN represent the extended features of intra sentence in the previous and latter sentence, respectively.
Table 13. Results of event argument extraction under different methods.
Table 13. Results of event argument extraction under different methods.
TasksEA RecognitionEA Classification
Methods PRF1PRF1
Word2Vec + BiLSTM + CRF45.251.348.141.348.644.7
ELMO + BiLSTM + CRF52.757.454.951.657.254.3
BERT + BiLSTM + CRF53.557.855.652.657.154.8
BERT + BiLSTM51.054.752.848.953.150.9
Table 14. Results of event argument extraction under different feature combinations.
Table 14. Results of event argument extraction under different feature combinations.
TasksEA RecognitionEA Classification
Methods PRF1PRF1
D54.258.356.253.157.655.3
DC57.863.460.556.259.457.8
DY54.557.956.152.757.154.8
DCY55.158.056.553.658.255.8
DCT60.562.661.558.560.159.3
DCTY58.759.859.255.657.356.4
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhai, P.; Wang, C.; Fang, Y. LSLSD: Fusion Long Short-Level Semantic Dependency of Chinese EMRs for Event Extraction. Appl. Sci. 2021, 11, 7237. https://0-doi-org.brum.beds.ac.uk/10.3390/app11167237

AMA Style

Zhai P, Wang C, Fang Y. LSLSD: Fusion Long Short-Level Semantic Dependency of Chinese EMRs for Event Extraction. Applied Sciences. 2021; 11(16):7237. https://0-doi-org.brum.beds.ac.uk/10.3390/app11167237

Chicago/Turabian Style

Zhai, Pengjun, Chen Wang, and Yu Fang. 2021. "LSLSD: Fusion Long Short-Level Semantic Dependency of Chinese EMRs for Event Extraction" Applied Sciences 11, no. 16: 7237. https://0-doi-org.brum.beds.ac.uk/10.3390/app11167237

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop