Next Article in Journal
A Group Resident Daily Load Forecasting Method Fusing Self-Attention Mechanism Based on Load Clustering
Next Article in Special Issue
Low-Resource Neural Machine Translation Improvement Using Source-Side Monolingual Data
Previous Article in Journal
The Safe Passage Redundancy Analysis of Airport Taxiway Bridge Based on Aircraft Load Fatigue Accumulation
Previous Article in Special Issue
Ensemble-NQG-T5: Ensemble Neural Question Generation Model Based on Text-to-Text Transfer Transformer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Span-Based Fine-Grained Entity-Relation Extraction via Sub-Prompts Combination

1
Key Laboratory of Trustworthy Distributed Computing and Service (BUPT), Ministry of Education, Beijing University of Posts and Telecommunications, Beijing 100876, China
2
School of Cyberspace Security (BUPT), Beijing University of Posts and Telecommunications, Beijing 100876, China
*
Author to whom correspondence should be addressed.
Submission received: 5 December 2022 / Revised: 10 January 2023 / Accepted: 13 January 2023 / Published: 15 January 2023
(This article belongs to the Special Issue Natural Language Processing (NLP) and Applications)

Abstract

:
With the development of information extraction technology, a variety of entity-relation extraction paradigms have been formed. However, approaches guided by these existing paradigms suffer from insufficient information fusion and too coarse extraction granularity, leading to difficulties extracting all triples in a sentence. Moreover, the joint entity-relation extraction model cannot easily adapt to the relation extraction task. Therefore, we need to design more fine-grained and flexible extraction methods. In this paper, we propose a new extraction paradigm based on existing paradigms. Then, based on it, we propose SSPC, a method for Span-based Fine-Grained Entity-Relation Extraction via Sub-Prompts Combination. SSPC first decomposes the task into three sub-tasks, namely S , R Extraction, R , O Extraction and S , R , O Classification and then uses prompt tuning to fully integrate entity and relation information in each part. This fine-grained extraction framework makes the model easier to adapt to other similar tasks. We conduct experiments on joint entity-relation extraction and relation extraction, respectively. The experimental results show that our model outperforms previous methods and achieves state-of-the-art results on ADE, TACRED, and TACREV.

1. Introduction

Information extraction is an important task in natural language processing, which aims to extract structured information from unstructured text. This task consists of two critical subtasks: Named Entity Recognition (NER) and Relation Extraction (RE). The former identifies various entities in a sentence, and the latter extracts whether and what relations exist between any two entities in a sentence. To extract complete structured information ( S u b j e c t , R e l a t i o n , O b j e c t , i.e., S , R , O ), researchers usually handle these two complementary tasks simultaneously. So far, the resolution process has gone through three phases, evolving into three kinds of extraction paradigms. The two early paradigms can be formulated as “ S , O [ R ] ” ( [ ] represents a model) [1,2,3] and “ [ S , O ] [ R ] ” [4,5,6,7]. However, neither paradigm considers the connection between entities and relations, leading to poor extraction results. To address this problem, a lot of current work has focused on accomplishing both tasks in a single model, which can be called Entity and Relation Extraction (ERE). Thus, the third type of extraction paradigm evolved. Depending on what to extract first, it is further divided into the following specific forms, i.e., P 1 : [ S , O R ] , P 2 : [ S R , O ] , and P 3 : [ R S , O ] [8,9,10,11,12,13,14,15].
However, the third paradigm still has some drawbacks. Under the guidance of these existing paradigms, there are some problems in the extraction process, such as insufficient information fusion and coarse extraction granularity. Specifically, there is only a one-step information connection between entities and relations in the extraction process. P 1 and P 2 first extract the entities in the sentence and then transfer the information of the entities to the relation extraction part. P 3 first extracts the relations in the sentence and then transfers the information of the relations to the entity extraction part. Nevertheless, in the first step of extraction, there is a lack of information supplement from relation information to entity extraction or entity information to relation extraction. Take the sentence “Neutropenia and agranulocytosis are risks known to occur with phenothiazines and clozapine.” in Figure 1 as an example. P 1 and P 2 are not guided by any relation information when extracting “Neutropenia”, “agranulocytosis”, “phenothiazines” and “clozapine” in the first step, even if the pre-defined target relation “Adverse-Effect” has been given in advance. The “Adverse-Effect” means “The effect was caused by the drug.”. The combination of “effect”, “drug”, and “was caused by” generates a large amount of information, including sequential information, causal information, and semantic information, which is very crucial for entity extraction. So, there is no reason to ignore this information, and we need to find a way to integrate it into the first step of the extraction process. Another issue is that the difference between relation extraction and joint entity and relation extraction is whether the entities in the sentence have been annotated. Therefore, a flexible ERE approach should ensure that it can quickly adapt to RE.
To address the challenges above, we propose a new extraction paradigm, which can be described as [ S , R & R , O S , R , O ] . It can not only extract entities but also extract more fine-grained elements under the guidance of relation: S , R and R , O . As the intermediate extraction result, these elements can form triples through logical combination and specific judgment as the final joint extraction result. To implement it, we propose SSPC, a model for Span-based Fine-Grained Entity-Relation Extraction with Sub-Prompts Combination. We introduce the approach of prompt tuning into the model. A typical prompt consists of a template and a set of label words. This form ingeniously conforms to our idea of integrating entity and relation information as long as we convert the relation information into a prompt. Therefore, we first design a set of prompts before extraction, including the S-R prompt, R-O prompt, and S-R-O prompt, and then begin the extraction process. Firstly, traverse all spans of the sentence, and extract a set of possible subjects for each relation through the S-R prompt; Secondly, traverse all spans of the sentence, and extract a set of possible objects for each relation through the R-O prompt; Thirdly, combine the above intermediate results through a logic rule to form a set of undetermined triples, and filter each element in the set through the S-R-O prompt and relation classifier to identify the triples that indeed exist in the sentence. SSPC can be easily applied to the relation extraction task as well. Since the entities in the sentence are already annotated, it is not necessary to traverse all spans of the sentence but place the annotated entity in the corresponding position of the template. After the same steps, we can achieve relation extraction.
  • We propose a new extraction paradigm. It solves the coarse-grained problem of the existing paradigm. A method guided by it can fully integrate entity and relation information in the extraction process to solve the problem of insufficient information fusion.
  • We adopt the new extraction paradigm and propose SSPC, a model for Span-based Fine-Grained Entity-Relation Extraction with Sub-Prompts Combination. Moreover, this model can also be easily adapted to the relation extraction task. SSPC first performs the S , R extraction and then performs the R , O extraction, and finally combines intermediate results through a logic rule to determine whether they are triples S , R , O that are indeed contained in the sentence.
  • We test our model on ADE, TACRED, and TACREV, and the results show that SSPC can significantly and consistently outperform existing state-of-the-art baselines.

2. Related Work

2.1. Joint Entity and Relation Extraction

The research direction of the joint entity and relation extraction is promising because it can solve the shortcomings of the previous methods, which ignore the interaction between sub-tasks and are affected by error propagation [16,17,18,19,20].
Under the guidance of the P 1 paradigm, Zeng et al. [8] proposed an end-to-end model based on sequence-to-sequence learning with copy mechanism, which can jointly extract relational facts from sentences of any of these classes. Both refs. [9,10] proposed a span-based joint entity and relation extraction model. The first work showed that with strong negative sampling, span filtering, and a localized context representation, a search over all spans in an input sentence becomes feasible for joint entity and relation extraction. The second work applied MLP attention to capture span-specific features aiming to obtain semantic rich span representation and calculated task-specific contextual representations with attention architecture to further reinforce span and relation representations. Unlike previous work based on BIO labels, these two approaches can identify overlapping entities and inspire us to think about the feasibility of the span-based approach.
Under the guidance of the P 2 paradigm, Li et al. [11] proposed a multi-turn question answering paradigm for the task of entity-relation extraction and Zhao et al. [12] provided an effective solution based on the machine reading comprehension models. These two models cast the entity-relation extraction to a multi-turn question answering (QA) and machine reading comprehension (MRC) task, enlightening us that joint extraction should adapt to more extraction paradigms and approaches.
Under the guidance of the P 3 paradigm, Chen et al. [13] proposed Patti, a novel pattern-first pipeline perspective for entity-relation extraction, which applied a Machine Reading Comprehension (MRC)-based framework to characterize entities and relations better. It alleviated entity redundancy and the entity overlap problem. Moreover, Takanobu et al. [14] applied a hierarchical reinforcement learning framework to enhance the interaction between entity mentions and relation types. Xie et al. [15] proposed a pipeline approach that first performs sentence classification with relational labels and then extracts the subjects and objects.
Recently proposed, the seq2seq model not only performs well in language generation but also in NLU tasks. Instead of tackling the joint extraction by training task-specific discriminative classifiers, Paolini et al. [21] framed it as a translation task between augmented natural languages, from which the task-relevant information can be easily extracted. Cabot et al. [22] performed relation extraction by representing triples as text sequences.
After analysis, under the guidance of existing paradigms, there is only a one-step information connection between entities and relations in the extraction process, resulting in insufficient information fusion. Moreover, these models are challenging to adapt to the relation extraction task.

2.2. Prompt Tuning

Recently, with the development of pre-trained language models, such as GPT [23], BERT [24], RoBERTa [25], T5 [26] and GPT-3 [27], the usage has also evolved from fine-tuning to prompt tuning. Prompt tuning has been widely used in various natural language processing tasks. Schick et al. [28] proposed an approach, PET, which consists of defining pairs of cloze question patterns and verbalizers that help leverage the knowledge contained within pre-trained language models for downstream tasks. It has achieved outstanding performance in widespread NLP tasks [29,30,31], especially in information extraction tasks [3,32,33,34,35]. Cui et al. [32] proposed a template-based method for NER, treating NER as a language model ranking problem in a sequence-to-sequence framework, where original sentences and statement templates filled by candidate named entity span are regarded as the source sequence and the target sequence, respectively. Lee et al. [33] proposed demonstration-based learning, a simple-yet-effective way to incorporate automatically constructed auxiliary supervision. Instead of reformatting the NER task into the cloze-style template, they augment the original input instances by appending automatically created task demonstrations. Ma et al. [34] proposed a template-free prompt tuning method, EntLM, for few-shot NER. In this way, not only the complicated template-based methods can be discarded, but also the few-shot performance can be boosted since the model objective reduces the gap between pretraining and fine-tuning. These works demonstrate the diversity of prompt learning in information extraction tasks.
On the relation extraction task, refs. [3,35] give us the best inspiration. The first work proposed prompt tuning with rules for many-class text classification and applied logic rules to construct prompts with several sub-prompts. In this way, it can encode prior knowledge of each class into prompt tuning. Although this work is defined on the relation extraction task, we apply it to our framework through transformation. The second work incorporated knowledge among relation labels into prompt-tuning for relation extraction and proposes a knowledge-aware prompt-tuning approach with synergistic optimization. This approach is flexible. In the experiment, we also use this approach as a robust baseline.

3. Methodology

The overall framework of our approach is shown in Figure 2, in which different pre-trained language models (PLMs) are selected as the core according to the dataset characteristics.
Given a sentence:
S e n t = { w 1 , w 2 , w 3 , . . . , w n } ,
we will focus on all spans in the sentence:
S p a n = { w i , w i + 1 , w i + 2 , . . . , w i + l } ,
where w denotes tokens, n denotes the sentence length and l denotes the span width. In this way, we will extract which span is the subject (S) of a relation (R) ( S , R ), which span is the object (O) of a relation (R) ( R , O ), and which two elements ( S , R and R , O ) can form a triple ( S , R , O ).
For example, the relation “Adverse-Effect” in the ADE dataset means “The effect was caused by the drug.”. This explanation can guide joint entity and relation extraction and decompose the extraction process into:
  • If s p a n 1 is of type “Effect” and the sentence indicates that it was caused by a type of “Drug”, then s p a n 1 can be the subject of the relation:
    s p a n 1 , A d v e r s e E f f e c t .
  • If s p a n 2 is of type “Drug” and the sentence indicates that it resulted in an “Effect”, then s p a n 2 can be the object of the relation:
    A d v e r s e E f f e c t , s p a n 2 .
  • If the sentence expresses that s p a n 1 and s p a n 2 have a “Cause” relation in the sentence, they can form a triple:
    s p a n 1 , A d v e r s e E f f e c t , s p a n 2 .
Based on the above considerations, we propose SSPC, an approach for Span-based Fine-Grained Entity-Relation Extraction with Sub-Prompts Combination. In the following sections, we describe each part of the framework in detail.

3.1. Sub-Prompts Design

In order to achieve the extraction and classification tasks of the three parts and fully integrate the information of entities and relations, we naturally introduce the approach of prompt tuning. We design a set of prompts, including the S-R prompt, R-O prompt, and S-R-O prompt. Similar to the conventional prompt setting, each sub-prompt consists of a template and a set of label words.
We still take the ADE dataset as an example. Firstly, in preparation, we add the “no_relation” relation, which means “The nothing is irrelevant to the nothing”. Then, the S-R Template, R-O Template, and S-R-O Template can be formalized as:
S R T e m p l a t e s p a n 1 = T h e [ M A S K ] 1 s p a n 1 [ M A S K ] 2 t h e [ M A S K ] 3 .
R O T e m p l a t e s p a n 2 = T h e [ M A S K ] 1 [ M A S K ] 2 t h e [ M A S K ] 3 s p a n 2 .
S R O T e m p l a t e s p a n 1 , s p a n 2 = T h e [ M A S K ] 1 s p a n 1 [ M A S K ] 2 t h e [ M A S K ] 3 s p a n 2 .
The position of [ M A S K ] 1 can tell us the property of s p a n 1 ; the position of [ M A S K ] 3 can tell us the property of s p a n 2 ; the position of [ M A S K ] 2 can tell us the action of s p a n 1 or s p a n 2 and the interaction of s p a n 1 and s p a n 2 in the sentence. Therefore, to ensure grammatical correctness and semantic integrity, we design label words for [ M A S K ] 1 , [ M A S K ] 2 and [ M A S K ] 3 , respectively, according to the characteristics of ADE dataset. The aggregated sets of label words are given as:
L a b e l s _ M A S K 1 = e f f e c t , n o t h i n g
L a b e l s _ M A S K 2 = w a s c a u s e d b y , i s i r r e l e v a n t t o
L a b e l s _ M A S K 3 = d r u g , n o t h i n g
Note that the set of label words of [ M A S K ] 1 is the same in the three templates, as are [ M A S K ] 2 and [ M A S K ] 3 . These three prompts will be applied to S , R Extraction, R , O Extraction, and S , R , O Classification, respectively.

3.2. S , R Extraction

Given a sentence, we will traverse all spans of the sentence. We first put the current span { w i , w i + 1 , w i + 2 , , w i + l } in the corresponding position of the S-R template and splice the sentence and the template to form
{ [ C L S ] , S e n t e n c e , T e m p l a t e , [ S E P ] } ,
and use the PLM to encode all tokens of the input sequence into responding vectors. Then, we take out the responding vectors h MASK 1 , h MASK 2 , and h MASK 3 of M A S K 1 , M A S K 2 , and M A S K 3 and calculate the possibility of each label word in the corresponding labels set to fill in the position:
p ( [ M A S K ] = w ) = e x p ( h [ MASK ] · w ) w ˜ L a b e l s _ [ M A S K ] h [ MASK ] · w ˜ )
where w is the embedding of the label word w in the PLM. Finally, we calculate the possibility of this span as the subject S of a relation R ^ :
p ( s p a n , R ^ ) = p ( [ M A S K ] 1 = l a b e l s _ [ M A S K ] 1 [ i ] ) + p ( [ M A S K ] 2 = l a b e l s _ [ M A S K ] 2 [ j ] ) + p ( [ M A S K ] 3 = l a b e l s _ [ M A S K ] 3 [ k ] )
where i is the index of the word of the R ^ at the [ M A S K ] 1 position in the L a b e l s _ [ M A S K ] 1 ; j is the index of the word of the R ^ at the [ M A S K ] 2 position in the L a b e l s _ [ M A S K ] 2 ; k is the index of the word of the R ^ at the [ M A S K ] 3 position in the L a b e l s _ [ M A S K ] 3 .
Thus, after traversing all spans, we will extract a set of possible subjects for each relation.

3.3. R , O Extraction

Similar to Section 3.2, we calculate the possibility of this span as the object O of a relation R ^ :
p ( R ^ , s p a n ) = p ( [ M A S K ] 1 = l a b e l s _ [ M A S K ] 1 [ i ] ) + p ( [ M A S K ] 2 = l a b e l s _ [ M A S K ] 2 [ j ] ) + p ( [ M A S K ] 3 = l a b e l s _ [ M A S K ] 3 [ k ] )
where i is the index of the word of the R ^ at the [ M A S K ] 1 position in the L a b e l s _ [ M A S K ] 1 ; j is the index of the word of the R ^ at the [ M A S K ] 2 position in the L a b e l s _ [ M A S K ] 2 ; k is the index of the word of the R ^ at the [ M A S K ] 3 position in the L a b e l s _ [ M A S K ] 3 .
Thus, after traversing all spans, we will extract a set of possible objects for each relation.

3.4. S , R , O Classification

After calculation and reasoning in Section 3.2 and Section 3.3, we obtain all the S , R ^ elements and R ^ , O elements in the sentence as intermediate results. We apply a simple strategy of using a logic rule and directly concatenate S , R ^ and R ^ , O of the same R ^ to form S , R ^ , O as the element in the set of pending triples. Then, we filter each element through the S-R-O prompt and relation classifier to identify the triples that really exist in the sentence.
On the one hand, we put S and O in the positions of s p a n 1 and s p a n 2 in the S-R-O template. Similar to Section 3.2 and Section 3.3, we calculate the possibility of S and O as the subject and object of a relation R ^ , respectively:
p ( < s p a n 1 , R ^ , s p a n 2 > ) = p ( [ M A S K ] 1 = l a b e l s _ [ M A S K ] 1 [ i ] ) + p ( [ M A S K ] 2 = l a b e l s _ [ M A S K ] 2 [ j ] ) + p ( [ M A S K ] 3 = l a b e l s _ [ M A S K ] 3 [ k ] )
where i is the index of the word of the R ^ at the [ M A S K ] 1 position in the L a b e l s _ [ M A S K ] 1 ; j is the index of the word of the R ^ at the [ M A S K ] 2 position in the L a b e l s _ [ M A S K ] 2 ; k is the index of the word of the R ^ at the [ M A S K ] 3 position in the L a b e l s _ [ M A S K ] 3 .
Thus, we will give the first relation prediction R pred 1 about this entity pair s p a n 1 and s p a n 2 .
On the other hand, we fuse the information of S and O and input them into the relation classifier to predict whether they have a relation and what kind of relation they have. The input consists of five parts:
  • The embedding of s p a n 1 . All the token embeddings are combined using a fusion, f ( e i , e i + 1 , e i + 2 , , e i + l ) . Regarding the fusion function f, we choose max-pooling, obtaining the s p a n 1 ’s representation e ( span 1 ) .
  • The size embedding of s p a n 1 . Given the span size l, we look-up a size embedding from a dedicated embedding matrix, obtaining the s p a n 1 ’s size representation s ( span 1 size ) . These size embeddings are learned by backpropagation.
  • The embedding of context. Obviously, words from the context are essential indicators of the expressed relation. We use a more localized context drawn from the direct surrounding of the spans: Given the span ranging from the end of the s p a n 1 to the beginning of the s p a n 2 , we combine its embeddings by max-pooling, obtaining a context representation c ( span 1 , span 2 ) . If the range is empty (e.g., in case of overlapping entities), we set c ( span 1 , span 2 ) = 0 .
  • The embedding of s p a n 2 . Similar to the s p a n 1 ’s embedding, we obtain the s p a n 2 ’s representation e ( span 2 ) .
  • The size embedding of s p a n 2 . Similar to the s p a n 2 ’s size embedding, we obtain the s p a n 2 ’s size representation s ( span 2 size ) .
The final input to the relation classifier is (whereas ∘ denotes concatenation):
X ( s p a n 1 , s p a n 2 ) = e ( s p a n 1 ) s ( s p a n 1 s i z e ) c ( s p a n 1 , s p a n 2 ) e ( s p a n 2 ) s ( s p a n 2 s i z e )
Then, it was passed through a single-layer classifier:
y ^ ( s p a n 1 , s p a n 2 ) = σ ( w · X ( s p a n 1 , s p a n 2 ) + b )
where σ denotes a softmax function or a sigmoid function. The highest response in the layer indicates that the corresponding relation R pred 2 holds between s p a n 1 and s p a n 2 .
Finally, if R ^ , R p r e d 1 , and R p r e d 2 are equal, we will choose this S , R ^ , O as one of the results of the entity relation extraction of the sentence.
Thus, after traversing all pending triples, we will obtain a set of actual triples in the sentence, which completes the entire entity relation extraction process.

3.5. Joint Training

Our proposed framework consists of three parts, which are trained jointly. We define the loss function of the overall framework:
L = L < S , R > + L < R , O > + L < S , R , O >
The first term is defined on the < S , R > Extraction. It includes two parts, one is the sum of cross-entropy loss of each masked position label, the other is the cross-entropy loss of span label. The second term is defined on the < R , O > Extraction, the loss is the same as the first term. The third term is defined on the < S , R , O > Extraction. It includes two parts, one is the same as the first term, the other is the cross-entropy loss of the relation label of s p a n 1 and s p a n 2 .
For all these three parts, we construct many within-sentence negative examples to train:
  • For the S , R Extraction and R , O Extraction, We construct n within-sentence negative examples (“no_relation”) by
    • replacing the subject and object of the R (the subject acts as the object, and the object acts as the subject),
    • blurring the subject or object’s boundary of the R (expanding or narrowing the starting and ending positions of a span), and
    • generating another unrelated span as the subject or the object of the R.
    These negative examples are combined with the existing positive examples to form a training set.
  • For the S , R , O Extraction, We construct n within-sentence negative examples (“no_relation”) by
    • replacing the subject and object of a single triple,
    • cross replacing the subjects and objects of multiple triples,
    • blurring the subject or object’s boundary of a single triple, and
    • generating other unrelated spans as subjects and objects to form triples.
    These negative examples are combined with the existing positive examples to form a training set.
  • For example, given the sentence, “Adriamycin—induced cardiomyopathy aggravated by cis—platinum nephrotoxicity requiring dialysis.”, we will construct negative samples
    A d r i a m y c i n , n o _ r e l a t i o n , p l a t i n u m n e p h r o t o x i c i t y , n o _ r e l a t i o n , a g g r a v a t e d , n o _ r e l a t i o n , etc .
    for the S , R Extraction;
    n o _ r e l a t i o n , c a r d i o m y o p a t h y , n o _ r e l a t i o n , p l a t i n u m , n o _ r e l a t i o n , d i a l y s i s , etc .
    for the R , O Extraction;
    A d r i a m y c i n , n o _ r e l a t i o n , c a r d i o m y o p a t h y , c a r d i o m y o p a t h y , n o _ r e l a t i o n , c i s p l a t i n u m , n e p h r o t o x i c i t y , n o _ r e l a t i o n , p l a t i n u m , i n d u c e d , n o _ r e l a t i o n , d i a l y s i s , etc .
    for the S , R , O Extraction.

3.6. Adapt to the Relation Extraction Task

The difference between relation extraction and joint entity relation extraction is whether entities in the sentence have been annotated. Therefore, a flexible ERE approach should ensure that it can quickly adapt to the RE. Our proposed SSPC guarantees this. Since the entities in the sentence are already annotated, it is not necessary to traverse all spans of the sentence but place the annotated entity in the corresponding position of the template. After the same steps, we can achieve relation extraction.

4. Experiments

4.1. Datasets

Our experiments are conducted on three widely used benchmarks: ADE, TACRED and TACREV.
  • ADE: The ADE dataset [36] is about drugs and their adverse effects. A sentence contains entities of drug type and adverse effect type and expresses the corresponding relation between them. It consists of 4272 sentences, of which 1695 contain overlapping phenomena. As shown in Table 1, there are triples in the ADE dataset. It can be seen that even if only one relation is involved, entities and relations have become complex, mainly including multiple triples (1), subjects and R , O many-to-one (2), S , R and objects one-to-many (3), and subjects and objects crossing (4). As in previous work, we conduct 10-fold cross-validation. The core pre-trained language model is “microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext”; the maximum length of a sentence is set to 128; the maximum length of a span is set to 5; the number of within-sentence negative samples is set to 70. The experiments are conducted on an NVIDIA Geforce RTX 3090 24GB GPU. Other hyperparameters are shown in Table 2.
  • TACRED and TACREV: The TACRED datatset [37] is a sizeable relation extraction dataset with 106,264 examples (68,124 for training, 22,631 for validation, and 15,509 for testing). These examples are created by combining available human annotations from the TAC KBP challenges and crowdsourcing. It contains 42 relation types, such as “per:parents”, “org:website”, and “no_relation”. The TACREV dataset [38] is built based on the original TACRED dataset. It finds out and corrects the errors in the original development set and test set of TACRED, while the training set is left intact. To experiment, we set entity types according to relation types, including “person”, “organization”, “religion”, “country”, “state”, “city”, “title”, “number”, “URL”, “event”, and “date”. For few-shot learning, consistent with the previous work, we sample K training and K validation instances per class from the original training and development set and evaluate models on the original test set. We set K from {8, 16, 32}, respectively. The core pre-trained language model is “roberta-large”; the maximum length of a sentence is set to 256; the maximum length of a span is set to 5. The experiments are conducted on an NVIDIA Geforce RTX 3090 24GB GPU. Other hyperparameters are shown in Table 2.

4.2. Comparison with the State-of-the-Art Model of the ADE Dataset

In this part, we compare SSPC with other joint entity relation extraction models, including the state-of-the-art model:
  • Natural Language Understanding (NLU) models, including “CNN + Global features” proposed by [16], “BiLSTM + SDP” proposed by [17], “Multi-head” proposed by [18], “Multi-head + AT” proposed by [19], “Relation-Metric” proposed by [20], “SpERT” proposed by [10], “SPAN” proposed by [9].
    “SpERT” and “SPAN” are both span-based joint entity and relation extraction models. “SpERT” is a span-based joint entity and relation extraction with transformer pre-training model; “SPAN” is a span-based joint entity and relation extraction with attention-based span-specific and contextual semantic representations model and is the current span-based state-of-the-art model on the ADE dataset.
  • Natural Language Generation (NLG) models, including “TANL” proposed by [21] and “REBEL” proposed by [22].
    Recently, the seq2seq model not only performs well in language generation but also in NLU tasks. “TANL” is generative as it translates from an input to an output in augmented natural languages. “REBEL” is a seq2seq model based on BART that performs end-to-end relation extraction and is the current state-of-the-art model on the ADE dataset.
Table 3 shows our experimental results. The second column is the names of the baseline models. On the ADE dataset, we use standard P r e c i s i o n , R e c a l l , and F 1 for evaluation. Furthermore, when S and O’s boundaries are correct, we treat a S , R , O as correct. We report the mean performance over 10-fold cross-validation to compare results to established work. From the results, we can see that SSPC consistently outperforms both NLU and NLG SOTA models. Compared with SPAN in entity recognition, P r e c i s i o n increases by 0.78% and F 1 increases by 0.4%. Compared with SPAN in relation extraction, P r e c i s i o n increases by 1.23%, R e c a l l increases by 0.93%, and F 1 increases by 1.06%. Compared with REBEL in relation extraction, F 1 increases by 0.1%. We owe these performance increases to our new extraction paradigm and fine-grained extraction framework.

4.3. Comparison with the State-of-the-Art Model of Relation Extraction

In this part, we compare SSPC with several recent relation extraction models using prompts. Moreover, we validate the model’s capabilities in two settings, including standard supervised training and few-shot learning.
  • ENT MARKER and TYP MARKER: [39] uses prompt tuning for relation extraction. ENT MARKER injects special symbols to index the positions of entities. It is similar to prompting by introducing extra serial information to indicate the position of special tokens, i.e., named entities. TYP MARKER additionally introduces the type information of entities. It could be regarded as a type of template for prompts but requires additional annotation of type information.
  • PTR: [3] proposes prompt tuning with rules (PTR) for many-class text classification and applies logic rules to construct prompts with several sub-prompts.
  • KnowPrompt: [35] incorporates knowledge among relation labels into prompt-tuning for relation extraction and proposes a knowledge-aware prompt-tuning approach with synergistic optimization. It is the current state-of-the-art model on TACTED and TACREV in few-shot learning scenarios.
Table 4 shows our experimental results. The first column is the names of the baseline models. We use micro-averaged F 1 (excluding “no_relation” type) for evaluation and treat a S , R , O as correct when S and O’s boundaries are correct. From the results, we can see that our model performs well. In the standard supervised setting, F 1 increases by 0.3% on the TACRED dataset and 0.7% on the TACREV dataset. In the few-shot setting, F 1 increases by 0.3% on the TACRED dataset and 1.6% on the TACREV dataset on average. Note that when K is set to 32, the effect is the best, and F 1 increases by 3.0%. The experimental results prove that SSPC has fully adapted to the relation extraction task and is very effective in few-shot learning scenarios. Without introducing additional knowledge, SSPC still outperforms other robust baseline models. We owe these performance increases to our sub-prompts combination method.

4.4. Ablation Study

As shown in Table 5, we conduct ablation experiments to further evaluate the contribution of each part of the “ S , R , O Classification ” in our proposed SSPC. “w/o (3-2)” denotes ablating the part of (3-2) in Figure 2, leaving only the part of (3-2) to filter the elements in the intermediate result set of the triple; “w/o (3-1)” denotes ablating the part of (3-1) in Figure 2, leaving only the part of (3-2) to filter the elements in the intermediate result set of the triple; “w/o (3-1) and (3-2)” denotes ablating the part of (3-1) and (3-2) in Figure 2, regarding all the elements in the intermediate result set as the actual triples in the sentence without filtering. we observe that both (3-1) and (3-2) are helpful for the joint extraction. When we ablate (3-1) and (3-2) simultaneously, R e c a l l increases significantly, and P r e c i s i o n decreases significantly, indicating that we need these two parts to filter intermediate results.

4.5. S , R and R , O Extraction Inspection

Since our model can extract more fine-grained elements S , R and R , O , we also observed their extraction results on the ADE dataset. The experimental results are shown in Table 6. The P r e c i s i o n of E f f e c t , A d v e r s e E f f e c t extraction reaches 85.27%, the R e c a l l reaches 86.72%, and the F 1 reaches 85.99%; the P r e c i s i o n of A d v e r s e E f f e c t , D r u g extraction reaches 96.05%, the R e c a l l reaches 95.92%, and the F 1 reaches 95.99%. By analyzing the results, we can make the following observations. Firstly, SSPC can not only complete the entity recognition task, but also achieve more fine-grained extraction by integrating entity and relation information, which is critical for subsequent tasks. Secondly, SSPC has a solid ability to capture D r u g -type entities. However, the extraction ability of E f f e c t -type entities is unsatisfactory. We speculate that the pre-trained model has more and better knowledge of drugs.

5. Conclusions

In this paper, we propose a new extraction paradigm. Based on it, we propose SSPC, an approach for Span-based Fine-Grained Entity-Relation Extraction via Sub-Prompts Combination. Experiments show that SSPC, under the guidance of the new paradigm, can comprehensively model the information fusion of entities and relations in the extraction process, extract all triples in sentences, and adapt to relation extraction. In future work, we will design more advanced model structures to extract information under the guidance of the new paradigm. We also plan to explore more advanced methods for stimulating knowledge in pre-trained language models.

Author Contributions

Conceptualization, N.Y.; Methodology, N.Y.; Formal analysis, N.Y.; Data curation, N.Y.; Writing—original draft, N.Y.; Writing—review & editing, N.Y., J.L. and Y.S.; Funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (U21B2020, U1936216).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zeng, D.; Liu, K.; Lai, S.; Zhou, G.; Zhao, J. Relation classification via convolutional deep neural network. In Proceedings of the COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers, Dublin, Ireland, 23–29 August 2014; pp. 2335–2344. [Google Scholar]
  2. Zhou, P.; Shi, W.; Tian, J.; Qi, Z.; Li, B.; Hao, H.; Xu, B. Attention-based bidirectional long short-term memory networks for relation classification. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 7–12 August 2016; pp. 207–212. [Google Scholar]
  3. Han, X.; Zhao, W.; Ding, N.; Liu, Z.; Sun, M. PTR: Prompt Tuning with Rules for Text Classification. AI Open 2022, 3, 182–192. [Google Scholar] [CrossRef]
  4. Miwa, M.; Bansal, M. End-to-End Relation Extraction using LSTMs on Sequences and Tree Structures. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 7–12 August 2016; pp. 1105–1116. [Google Scholar]
  5. Adel, H.; Schütze, H. Global Normalization of Convolutional Neural Networks for Joint Entity and Relation Classification. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; pp. 1723–1729. [Google Scholar]
  6. Katiyar, A.; Cardie, C. Going out on a limb: Joint extraction of entity mentions and relations without dependency trees. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vancouver, BC, Canada, 30 July–4 August 2017; pp. 917–928. [Google Scholar]
  7. Yamada, I.; Asai, A.; Shindo, H.; Takeda, H.; Matsumoto, Y. LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; pp. 6442–6454. [Google Scholar]
  8. Zeng, X.; Zeng, D.; He, S.; Liu, K.; Zhao, J. Extracting relational facts by an end-to-end neural model with copy mechanism. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia, 15–20 July 2018; pp. 506–514. [Google Scholar]
  9. Ji, B.; Yu, J.; Li, S.; Ma, J.; Wu, Q.; Tan, Y.; Liu, H. Span-based joint entity and relation extraction with attention-based span-specific and contextual semantic representations. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain, 8–13 December 2020; pp. 88–99. [Google Scholar]
  10. Eberts, M.; Ulges, A. Span-Based Joint Entity and Relation Extraction with Transformer Pre-Training. In Proceedings of the ECAI 2020, Santiago de Compostela, Spain, 29 August–8 September 2020; pp. 2006–2013. [Google Scholar]
  11. Li, X.; Yin, F.; Sun, Z.; Li, X.; Yuan, A.; Chai, D.; Zhou, M.; Li, J. Entity-Relation Extraction as Multi-Turn Question Answering. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 1340–1350. [Google Scholar]
  12. Zhao, T.; Yan, Z.; Cao, Y.; Li, Z. Asking effective and diverse questions: A machine reading comprehension based framework for joint entity-relation extraction. In Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence, Yokohama, Japan, 7–15 January 2021; pp. 3948–3954. [Google Scholar]
  13. Chen, Z.; Guo, C. A pattern-first pipeline approach for entity and relation extraction. Neurocomputing 2022, 494, 182–191. [Google Scholar] [CrossRef]
  14. Takanobu, R.; Zhang, T.; Liu, J.; Huang, M. A hierarchical framework for relation extraction with reinforcement learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 7072–7079. [Google Scholar]
  15. Xie, C.; Liang, J.; Liu, J.; Huang, C.; Huang, W.; Xiao, Y. Revisiting the Negative Data of Distantly Supervised Relation Extraction. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Online, 5–6 August 2021; pp. 3572–3581. [Google Scholar]
  16. Li, F.; Zhang, Y.; Zhang, M.; Ji, D. Joint Models for Extracting Adverse Drug Events from Biomedical Text. In Proceedings of the IJCAI, New York, NY, USA, 9–15 July 2016; Volume 2016, pp. 2838–2844. [Google Scholar]
  17. Li, F.; Zhang, M.; Fu, G.; Ji, D. A neural joint model for entity and relation extraction from biomedical text. BMC Bioinform. 2017, 18, 1–11. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Bekoulis, G.; Deleu, J.; Demeester, T.; Develder, C. Joint entity recognition and relation extraction as a multi-head selection problem. Expert Syst. Appl. 2018, 114, 34–45. [Google Scholar] [CrossRef] [Green Version]
  19. Bekoulis, I.; Deleu, J.; Demeester, T.; Develder, C. Adversarial training for multi-context joint entity and relation extraction. In Proceedings of the EMNLP2018, the Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 1–7. [Google Scholar]
  20. Tran, T.; Kavuluru, R. Neural metric learning for fast end-to-end relation extraction. arXiv 2019, arXiv:1905.07458. [Google Scholar]
  21. Paolini, G.; Athiwaratkun, B.; Krone, J.; Ma, J.; Achille, A.; Anubhai, R.; dos Santos, C.N.; Xiang, B.; Soatto, S. Structured Prediction as Translation between Augmented Natural Languages. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  22. Cabot, P.L.H.; Navigli, R. REBEL: Relation extraction by end-to-end language generation. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2021, Punta Cana, Dominican Republic, 16–20 November 2021; pp. 2370–2381. [Google Scholar]
  23. Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving language understanding by generative pre-training. OpenAI 2018. Available online: https://www.cs.ubc.ca/~amuham01/LING530/papers/radford2018improving.pdf (accessed on 1 December 2022).
  24. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the NAACL-HLT, Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
  25. Liu, X.; Zheng, Y.; Du, Z.; Ding, M.; Qian, Y.; Yang, Z.; Tang, J. GPT understands, too. arXiv 2021, arXiv:2103.10385. [Google Scholar]
  26. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 2020, 21, 1–67. [Google Scholar]
  27. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  28. Schick, T.; Schütze, H. Exploiting cloze questions for few shot text classification and natural language inference. arXiv 2020, arXiv:2001.07676. [Google Scholar]
  29. Rubin, O.; Herzig, J.; Berant, J. Learning to retrieve prompts for in-context learning. arXiv 2021, arXiv:2112.08633. [Google Scholar]
  30. Li, J.; Tang, T.; Nie, J.Y.; Wen, J.R.; Zhao, W.X. Learning to Transfer Prompts for Text Generation. arXiv 2022, arXiv:2205.01543. [Google Scholar]
  31. Kasahara, T.; Kawahara, D.; Tung, N.; Li, S.; Shinzato, K.; Sato, T. Building a Personalized Dialogue System with Prompt-Tuning. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Student Research Workshop, Online, 10–15 July 2022; pp. 96–105. [Google Scholar]
  32. Cui, L.; Wu, Y.; Liu, J.; Yang, S.; Zhang, Y. Template-based named entity recognition using BART. arXiv 2021, arXiv:2106.01760. [Google Scholar]
  33. Lee, D.H.; Agarwal, M.; Kadakia, A.; Pujara, J.; Ren, X. Good examples make A faster learner: Simple demonstration-based learning for low-resource NER. arXiv 2021, arXiv:2110.08454. [Google Scholar]
  34. Ma, R.; Zhou, X.; Gui, T.; Tan, Y.; Zhang, Q.; Huang, X. Template-free prompt tuning for few-shot NER. arXiv 2021, arXiv:2109.13532. [Google Scholar]
  35. Chen, X.; Zhang, N.; Xie, X.; Deng, S.; Yao, Y.; Tan, C.; Huang, F.; Si, L.; Chen, H. Knowprompt: Knowledge-aware prompt-tuning with synergistic optimization for relation extraction. In Proceedings of the ACM Web Conference 2022, Lyon, France, 25–29 April 2022; pp. 2778–2788. [Google Scholar]
  36. Gurulingappa, H.; Rajput, A.M.; Roberts, A.; Fluck, J.; Hofmann-Apitius, M.; Toldo, L. Development of a benchmark corpus to support the automatic extraction of drug-related adverse effects from medical case reports. J. Biomed. Inform. 2012, 45, 885–892. [Google Scholar] [CrossRef] [PubMed]
  37. Zhang, Y.; Zhong, V.; Chen, D.; Angeli, G.; Manning, C.D. Position-aware attention and supervised data improve slot filling. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017. [Google Scholar]
  38. Alt, C.; Gabryszak, A.; Hennig, L. TACRED Revisited: A Thorough Evaluation of the TACRED Relation Extraction Task. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, online, 5–10 July 2020; pp. 1558–1569. [Google Scholar]
  39. Zhou, W.; Chen, M. An Improved Baseline for Sentence-level Relation Extraction. In Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing, Minneapolis, MN, USA, 20–21 November 2022; pp. 161–168. [Google Scholar]
Figure 1. The example is chosen from the ADE dataset, where subjects are marked in orange, objects are marked in blue, and relations are marked with red, green, purple, and black lines.
Figure 1. The example is chosen from the ADE dataset, where subjects are marked in orange, objects are marked in blue, and relations are marked with red, green, purple, and black lines.
Applsci 13 01159 g001
Figure 2. Our framework SSPC consists of three parts, S , R Extraction (1), R , O Extraction (2), and S , R , O Classification((3-1) and (3-2)).
Figure 2. Our framework SSPC consists of three parts, S , R Extraction (1), R , O Extraction (2), and S , R , O Classification((3-1) and (3-2)).
Applsci 13 01159 g002
Table 1. Triples in the ADE dataset, including multiple triples (1), subjects and R , O many-to-one (2), S , R and objects one-to-many (3), and subjects and objects crossing (4). The subjects are marked in orange, and the objects are marked in blue.
Table 1. Triples in the ADE dataset, including multiple triples (1), subjects and R , O many-to-one (2), S , R and objects one-to-many (3), and subjects and objects crossing (4). The subjects are marked in orange, and the objects are marked in blue.
1Adriamycin—induced cardiomyopathy aggravated by cis—platinum nephrotoxicity requiring dialysis.Applsci 13 01159 i001
{ <cardiomyopathy, Adverse-Effect, Adriamycin>,
  <nephrotoxicity, Adverse-Effect, cis—platinum> }
2Possible interaction between lopinavir/ritonavir and valproic Acid exacerbates  bipolar disorder.Applsci 13 01159 i002
{ <bipolar disorder, Adverse-Effect, lopinavir>,
  <bipolar disorder, Adverse-Effect, ritonavir>,
  <bipolar disorder, Adverse-Effect, valproic Acid> }
3Listeria brain abscess, Pneumocystis pneumonia and Kaposi ’s sarcoma
after  temozolomide .
Applsci 13 01159 i003
{ <Listeria brain abscess, Adverse-Effect, temozolomide>,
  <Pneumocystis pneumonia, Adverse-Effect, temozolomide>,
  <Kaposi ’s sarcoma, Adverse-Effect, temozolomide> }
4Neutropenia and agranulocytosis are risks known to occur with phenothiazines and clozapine .Applsci 13 01159 i004
{ <Neutropenia, Adverse-Effect, phenothiazines>,
  <Neutropenia, Adverse-Effect, clozapine>,
  <agranulocytosis, Adverse-Effect, phenothiazines>,
  <agranulocytosis, Adverse-Effect, clozapine> }
Table 2. Hyperparameters for the different datasets. (Setting: few-shot = ).
Table 2. Hyperparameters for the different datasets. (Setting: few-shot = ).
Max EpochsLearning RateWarm-UpWeight DecayBatch SizeTime Per Epoch
ADE20 5 × 10 5 10%0.01146 min× 10
TACRED20 3 × 10 5 10%0.018102 min
TACREV20 3 × 10 5 10%0.018102 min
TACRED 50 3 × 10 5 10%0.0180.5–2 min
TACREV 50 3 × 10 5 10%0.0180.5–2 min
Table 3. Main experimental results on the ADE dataset. The best results are bold.
Table 3. Main experimental results on the ADE dataset. The best results are bold.
DatasetModelEntityRelation
PrecisionRecallF1PrecisionRecallF1
ADECNN + Global features [16]79.5079.6079.5064.0062.9063.40
BiLSTM + SDP [17]82.7086.7084.6067.5075.8071.40
Multi-head [18]84.7288.1686.4072.1077.2474.58
Multi-head + AT [19]--86.73--75.52
Relation-Metric [20]86.1688.0887.1177.3677.2577.29
SpERT [10]88.9989.5989.2877.7779.9678.84
SPAN [9]89.8891.3290.5979.5681.9380.73
TANL [21]--90.20--80.60
REBEL [22]-----81.70
SSPC (ours)90.6691.3290.9980.7982.8681.79
Table 4. Our experimental results on the datasets TACRED and TACREV. F 1 scores of various models with different sizes of training instances.The best results are bold.
Table 4. Our experimental results on the datasets TACRED and TACREV. F 1 scores of various models with different sizes of training instances.The best results are bold.
ModelTACREDTACREV
Standard SettingFew-Shot SettingStandard SettingFew-Shot Setting
All Data81632MeanAll Data81632Mean
ENT
MARKER [39]
69.427.031.331.930.179.827.431.232.030.2
TYP
MARKER [39]
71.028.932.032.431.180.827.631.232.030.3
PTR [3]72.428.130.732.130.381.428.731.432.430.8
KnowPrompt [35]72.432.035.436.534.682.432.133.134.733.3
SSPC (ours)72.732.734.937.034.983.132.734.237.734.9
Table 5. Ablations on the ADE dataset.
Table 5. Ablations on the ADE dataset.
MethodPrecisionRecallF1
Full80.7982.8681.79
w/o (3-2)68.3380.1873.78
w/o (3-1)73.8481.2577.37
w/o (3-1) and (3-2)60.4087.7971.56
Table 6. Detailed experimental results on the ADE dataset.
Table 6. Detailed experimental results on the ADE dataset.
K Effect , Adverse Effect Adverse Effect , Drug
PrecisionRecallF1PrecisionRecallF1
186.7585.6986.2294.9794.3894.67
284.4090.0587.1395.9796.7596.36
385.5787.3786.4696.9595.0195.97
485.5685.4185.4995.6695.0795.37
585.8787.0786.4795.9297.5396.72
684.8184.3984.6097.7394.2395.95
784.8887.2086.0295.6195.6195.61
882.7285.8984.2796.7595.2195.98
986.4488.0887.2596.3397.5296.92
1085.7186.0185.8694.6197.9396.24
Mean85.2786.7285.9996.0595.9295.99
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, N.; Liu, J.; Shi, Y. Span-Based Fine-Grained Entity-Relation Extraction via Sub-Prompts Combination. Appl. Sci. 2023, 13, 1159. https://0-doi-org.brum.beds.ac.uk/10.3390/app13021159

AMA Style

Yu N, Liu J, Shi Y. Span-Based Fine-Grained Entity-Relation Extraction via Sub-Prompts Combination. Applied Sciences. 2023; 13(2):1159. https://0-doi-org.brum.beds.ac.uk/10.3390/app13021159

Chicago/Turabian Style

Yu, Ning, Jianyi Liu, and Yu Shi. 2023. "Span-Based Fine-Grained Entity-Relation Extraction via Sub-Prompts Combination" Applied Sciences 13, no. 2: 1159. https://0-doi-org.brum.beds.ac.uk/10.3390/app13021159

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop