Revisiting the Negative Data of Distantly Supervised Relation Extraction
Abstract
Distantly supervision automatically generates plenty of training samples for relation extraction. However, it also incurs two major problems: noisy labels and imbalanced training data. Previous works focus more on reducing wrongly labeled relations (false positives) while few explore the missing relations that are caused by incompleteness of knowledge base (false negatives). Furthermore, the quantity of negative labels overwhelmingly surpasses the positive ones in previous problem formulations. In this paper, we first provide a thorough analysis of the above challenges caused by negative data. Next, we formulate the problem of relation extraction into as a positive unlabeled learning task to alleviate false negative problem. Thirdly, we propose a pipeline approach, dubbed ReRe, that performs sentence-level relation detection then subject/object extraction to achieve sample-efficient training. Experimental results show that the proposed method consistently outperforms existing approaches and remains excellent performance even learned with a large quantity of false positive samples.
1 Introduction
Relational extraction is a crucial step towards knowledge graph construction. It aims at identifying relational triples from a given sentence in the form of subject, relation, object, in short, . For example, given S1 in Figure 1, we hope to extract William Shakespeare, Birthplace, stratford-upon-Avon.
This task is usually modeled as a supervised learning problem and distant supervision Mintz et al. (2009) is utilized to acquire large-scale training data. The core idea is to obtain training data is through automatically labeling a sentence with existing relational triples from a knowledge base (KB). For example, given a triple and a sentence, if the sentence contains both and , distant supervision methods regard as a valid sample for the sentence. If no relational triples are applicable, the sentence is labeled as “NA”.

Despite the abundant training data obtained with distant supervision, nonnegligible errors also occur in the labels. There are two types of errors. In the first type, the labeled relation does not conform with the original meaning of sentence, and this type of error is referred to as false positive (FP). For example, in S2, The “Shakespeare spent the last few years of his life in Stratford-upon-Avon.” does not express the relation Birthplace, thus being a FP. Second, large amounts of relations in sentences are missing due to the incompleteness of KB, which is referred to as false negative (FN). For instance, in S3, “Buffett was born in 1930 in Omaha, Nebraska.” is wrongly labeled as NA since there is no relation (e.g., Birthplace) between Buffett and Omaha, Nebraska in the KB. Many efforts have been devoted to solving the FP problem, including pattern-based methods Jia et al. (2019), multi-instance learning methods Lin et al. (2016); Zeng et al. (2018a) and reinforcement learning methods Feng et al. (2018). Significant improvements have been made.
However, FN problem receives much less attention Min et al. (2013); Xu et al. (2013); Roller et al. (2015). To the best of our knowledge, none existing work with deep neural networks to solve this problem. We argue that this problem is fatal in practice since there are massive FN cases in datasets. For example, there exist at least 33% and 35% FN’s in NYT and SKE datasets, respectively. We will deeply analyze the problem in Section 2.1
Another huge problem for relation extraction is overwhelming negative labels. As is widely acknowledged, information extraction tasks are highly imbalanced in class labels Chowdhury and Lavelli (2012); Lin et al. (2018); Li et al. (2020). In particular, the negative labels account for most of the labels in relation extraction under almost any problem formulation, which makes relation extraction a hard machine learning problem. We systematically analyze this in Section 2.2.
In this paper, we address these challenges caused by negative data. Our main contribution can be summarized as follows.
-
•
We systematically compare the class distributions of different problem modeling and explain why first extract relation then entities, i.e., the third paradigm (P3) in Section 2.2, is superior to the others.
-
•
Based on the first point, we adopt P3 and propose a novel two-staged pipeline model dubbed ReRe. It first detects relation at sentence level and then extracts entities for a specific relation. We model the false negatives in relation extraction as “unlabeled positives” and propose a multi-label collective loss function.
-
•
Our empirical evaluations show that the proposed method consistently outperforms existing approaches, and achieves excellent performance even learned with a large quantity of false positive samples. We also provide two carefully annotated test sets aiming at reducing the false negatives of previous annotation, namely, NYT21 and SKE21, with 370 and 1150 samples, respectively.
2 Problem Analysis and Pilot Experiments
We use to denote a training instance, where is a sentence consisting of tokens labeled by a set of triples from the training set . For rigorous definition, can be viewed as an ordered set so that set operations can be applied. We assume , where is a finite set of all relations in . Other model/task-specific notations are defined after each problem formulation.
We now clarify some terms used in the introduction and title without formal definition. A negative sample refers to a triple . Negative label refers to the negative class label (e.g., usually “0” for binary classification), used for supervision with respect to task-specific models. Under different task formulation, the negative labels can be different. Negative data is a general term that includes both negative labels and negative samples. There are two kinds of false negatives. Relation-level false negative (S3 in Figure 1) refers to the situation where there exists , but is actually expressed by , and does not appear in other . Similarly, Entity-level false negative (S4 and S5 in Figure 1) means appears in other . Imbalanced class distribution means that the quantity of negative labels are much larger than positive ones.
2.1 Addressing the False Negatives
As shown in Table 1, the triples in NYT (SKE) datasets111Detailed description of datasets is in Sec. 5.1 labeled by Freebase222Bollacker et al. (2008) (BaiduBaike333https://baike.baidu.com/) is 88,253 (409,767), while the ones labeled by Wikidata444Vrandecic and Krötzsch (2014) (CN-DBPedia555 Xu et al. (2017)) are 58,135 (342,931). In other words, there exists massive FN matches if only labeled by one KB due to the incompleteness of KBs. Notably, we find that the FN rate is underestimated by previous researches Min et al. (2013); Xu et al. (2013), based on the manual evaluation of which there are 15%-35% FN matches. This discrepancy may be caused by human error. In specific, a volunteer may accidentally miss some triples. For example, as pointed out by Wei et al. (Wei et al., 2020, in Appendix C), the test set of NYT11 Hoffmann et al. (2011) missed lots of triples, especially when multiple relations occur in a same sentence, though labeled by human. That also provides an evidence that FN’s are harder to discover than FP’s.
NYT (English) | SKE (Chinese) | |||
# Sentence | 56,196 | 194,747 | ||
# Triples | # Rels | # Triples | # Rels | |
Original | 88,253 | 23 | 409,767 | 49 |
Re-labeled | 58,135 | 57 | 342,931 | 378 |
Intersection | 13,848 | 18 | 121,326 | 46 |
Union | 132,540 | 62 | 631,372 | 381 |
Original FNR | 0.33 | 0.35 | ||
Relabel FNR | 0.56 | 0.46 |
Paradigm | Theoretical | NYT10-HRL | NYT11-HRL | SKE | ||||
---|---|---|---|---|---|---|---|---|
=31,= 39.08 | =11, =39.46 | =51,= 54.67 | ||||||
then | – | – | 0.01421 | – | 0.00280 | – | 0.00494 | |
then | 0.0585 | 0.00093 | 0.0574 | 0.00257 | 0.0405 | 0.00067 | ||
then | 0.0390 | 0.00842 | 0.0826 | 0.00835 | 0.0344 | 0.00927 |
2.2 Addressing the Overwhelming Negative Labels
We point out that some of the previous paradigms designed for relation extraction aggravate the imbalance and lead to inefficient supervision. The mainstream approaches for relation extraction mainly fall into three paradigms depending on what to extract first.
-
P1
The first paradigm is a pipeline that begins with named entity recognition (NER) and then classifies each entity pair into different relations, i.e., [ then ]. It is adopted by many traditional approaches Mintz et al. (2009); Chan and Roth (2011); Zeng et al. (2014, 2015); Gormley et al. (2015); dos Santos et al. (2015); Lin et al. (2016).
-
P2
The second paradigm first detects all possible subjects in a sentence then identifies objects with respect to each relation, i.e., [ then ]. Specific implementation includes modeling relation extraction as multi-turn question answering Li et al. (2019), span tagging Yu et al. (2020) and cascaded binary tagging Wei et al. (2020).
-
P3
The third paradigm first perform sentence-level relation detection (cf. P1, which is at entity pair level.) then extract subjects and entities, i.e., [ then ]. This paradigm is largely unexplored. HRL Takanobu et al. (2019) is hitherto the only work to apply this paradigm based on our literature review.
We provide theoretical analysis of the output space and class prior with statistical support from three datasets (see Section 5.1 for description) of the three paradigms in Table 2. The second step of P1 can be compared with the first step of P3. Both of them find relation from a sentence (P1 with target entity pair given). Suppose a sentence contains entities666Below the same., the classifier has to decide relation from entity pairs, while in reality, relations are often sparse, i.e., . In other words, most entity pairs in P1 do not form valid relation, thus resulting in a low class prior. The situation is even worse when the sentence contains more entities, such as in NYT11-HRL. In addition, P1 is not sample efficient because the classifier will be trained/queried for the same sentences. For P2, we demonstrate with the problem formulation of CasRel Wei et al. (2020). The difference of the first-step class prior between P2 and P3 depends on the result of comparison between # relations and average sentence length (i.e., and ), which varies in different scenarios/domains. However, of P2 is extremely low, where a classifier has to decide from a space of . In contrast, P3 only has to decide from based on our task formulation (Section 3.1)
Other task formulations include jointly extracting the relation and entities Yu and Lam (2010); Li and Ji (2014); Miwa and Sasaki (2014); Gupta et al. (2016); Katiyar and Cardie (2017); Ren et al. (2017) and recently in the manner of sequence tagging Zheng et al. (2017), sequence-to-sequence learning Zeng et al. (2018b). In contrast to the aforementioned three paradigms, most of these methods actually provide an incomplete decision space that cannot handle all the situation of relation extraction, for example, the overlapping one Wei et al. (2020).
3 Solution Framework
3.1 Framework of ReRe
Given an instance from , the goal of training is to maximize the likelihood defined in Eq. (1). It is decomposed into two components by applying the definition of conditional probability, formulated in Eq. (2).
(1) | ||||
(2) |
where we use as a shorthand for , which means that occurs in the triple set w.r.t. ; Similarly, , stands for and , respectively. represents a subset of with a common relation . is an indicator function; when the condition happens. We denote by the model parameters.
Under this decomposition, relational triple extraction task is formulate into two subtasks: relation classification and entity extraction.
Relation Classification.
As is discussed, building relation classifier at entity-pair level will introduce excessive negative samples that forms a hard learning problem. Therefore, we alternatively model the relation classification at sentence level. Intuitively speaking, we hope that the model could capture what relation a sentence is expressing. We formalize it as a multi-label classification task.
(3) |
where is the probability that is expressing , the -th relation777 is parameterized by , omitted in the equation for clarity, below the same.. is the ground truth from the labeled data; is equivalent to while means the opposite.
Entity Extraction.
We then model entity extraction task. We observe that given the relation and context , it naturally forms a machine reading comprehension (MRC) task Chen (2018), where naturally fits into the paradigm of (query, context, answer). Particularly, the subjects and objects are continuous spans from , which falls into the category of span extraction. We adopt the boundary detection model with answer pointer Wang and Jiang (2017) as the output layer, which is widely used in MRC tasks. Formally, for a sentence of tokens,
(4) |
where represents the identifier of each pointer; refers to the probability of -th token being the start/end of the subject/object. is the ground truth from the training data; if occurs in at position from to , then and , otherwise ; the same applies for the objects.
3.2 Advantages
Our task formulation shows several advantages. By adopting P3 as paradigm, the first and foremost advantage of our solution is that it suffers less from the imbalanced classes (Section 2.2). Secondly, relation-level false negative is easy to recover. When modeled as a standard classification problem, many off-the-shelf methods on positive unlabeled learning can be leveraged. Thirdly, entity-level false negatives do not affect relation classification. Taking S5 in Figure 1 as an example, even though the Birthplace relation between William Swartz and Scranton is missing, the relation classifier can still capture the signal from the other sample with a same relation, i.e., Joe Biden, Birthplace, Scranton . Fourthly, this kind of modeling is easy to update with new relations without the need of retraining a model from bottom up. Only relation classifier needs to be redesigned, while entity extractor can be updated in an online manner without modifying the model structure. Last but not the least, relation classifier can be regarded as a pruning step when applied to practical tasks. Many existing methods treat relation extraction as question answering Li et al. (2019); Zhao et al. (2020). However, without first identifying the relation, they all need to iterate over all the possible relations and ask diverse questions. This results in extremely low efficiency where time consumed for predicting one sample may takes up to times larger than our method.
4 Our Model
The relational triple extraction task decomposed in Eq. (2) inspires us to design a two-staged pipeline, in which we first detect relation at sentence level and then extract subjects/objects for each relation. The overall architecture of ReRe is shown in Figure 2.

4.1 Sentence-level Relation Classifier
We first detect relation at sentence level. The input is a sequence of tokens and we denote by the output vector of the model, which aims to estimate in Eq. (3). We use BERT Devlin et al. (2019) for English and RoBERTa Liu et al. (2019) for Chinese, pre-trained language models with multi-layer bidirectional Transformer structure Vaswani et al. (2017), to encode the inputs888For convenience, we refer to the pre-trained Transformer as BERT hereinafter.. Specifically, the input sequence , which is fed into BERT for generating a token representation matrix , where is the hidden dimension defined by pre-trained Transformers. We take , which is the encoded vector of the first token [CLS], as the representation of the sentence. The final output of relation classification module is defined in Eq. (5).
(5) |
where and are trainable model parameters, representing weights and bias, respectively; denotes the sigmoid activation function.
4.2 Relation-specific Entity Extractor
After the relation detected at sentence-level, we extract subjects and objects for each candidate relation. We aim to estimate , of which each element corresponds to in Eq. (4), using a deep neural model. We take , the one-hot output vector of relation classifier, and generate query tokens using each of the detected relations (i.e., the “1”s in ). We are aware that many recent works Li et al. (2019); Zhao et al. (2020) have studied how to generate diverse queries for the given relation, which have the potential of achieving better performance. Nevertheless, that is beyond the scope of this paper. To keep things simple, we use the surface text of a relation as the query.
Next, the input sequence is constructed as . Like Section 4.1, we get the token representation matrix from BERT. The -th output pointer of entity extractor is defined by
(6) |
where is in accordance to Eq. (4); and are the corresponding parameters.
The final subject/object spans are generated by pairing the nearest / with /. Next, all subjects are paired to the nearest object. If multiple objects occur before the next subject appears, all subsequent objects will be paired with it until next subject occurs.
4.3 Multi-label Collective Loss function
In normal cases, the log-likelihood is taken as the learning objective. However, as is emphasized, there exists many false negative samples in the training data. Intuitively speaking, the negative labels cannot be simply considered as negative. Instead, a small portion of the negative labels should be considered as unlabeled positives and their influence towards the penalty should be eradicated. Therefore, we adopt cPU Xie et al. (2020), a collective loss function that is designed for positive unlabeled learning (PU learning). To briefly review, cPU considers the learning objective to be the correctness under a surrogate function,
(7) |
where they redefine the correctness function for PU learning as
(8) |
where is the ratio of false negative data (i.e., the unlabeled positive in the original paper).
We extend it to multi-label situation by embodying the original expectation at sample level. Due to the fact that class labels are highly imbalanced for our tasks, we introduce a class weight to downweight the positive penalty. For relation classifier,
(9) |
For entity extractor,
(10) |
In practice, we set , where is the ratio of false negative and is the class prior. Note that is not difficult to estimate for both relation classification and entity extraction task in practice. Besides various of methods in the PU learning du Plessis et al. (2015); Bekker and Davis (2018) for estimating it, an easy approximation is when , which happens to be the case for our tasks.
5 Experiments
5.1 Datasets
Our experiments are conducted on these four datasets101010We do not use WebNLG Gardent et al. (2017) and ACE04999https://catalog.ldc.upenn.edu/LDC2005T09 because these datasets are not automatically labeled by distant supervision. WebNLG is constructed by natural language generation with triples. ACE04 is manually labeled.. Some statistics of the datasets are provided in Table 1 and Table 2. In relation extraction, some datasets with the same names involve different preprocessing, which leads to unfair comparison. We briefly review all the datasets below and specify the operations to perform before applying each dataset.
-
•
NYT Riedel et al. (2010). NYT is the very first version among all the NYT-related datasets. It is based on the articles in New York Times111111https://www.nytimes.com/. We use the sentences from it to conduct the pilot experiment in Table 1. However, 1) it contains duplicate samples, e.g., 1504 in the training set; 2) It only labels the last word of an entity, which will mislead the evaluation results.
-
•
NYT10-HRL. & NYT11-HRL. These two datasets are based on NYT. The difference is that they both contain complete entity mentions. NYT10 Riedel et al. (2010) is the original one. and NYT11 Hoffmann et al. (2011) is a small version of NYT10 with 53,395 training samples and a manually labeled test set of 368 samples. We refer to them as NYT10-HRL and NYT11-HRL after preprocessed by HRL Takanobu et al. (2019) where they removed 1) training relation not appearing in the testing and 2) “NA” sentences. These two steps are almost adopted by all the compared methods. To compare fairly, we use this version in evaluations.
-
•
NYT21. We provide relabel version of the test set of NYT11-HRL. The test set of NYT11-HRL still have false negative problem. Most of the samples in the NYT11-HRL has only one relation. We manually added back the missing triples to the test set.
-
•
SKE2019/SKE21121212http://ai.baidu.com/broad/download?dataset=sked. SKE2019 is a dataset in Chinese published by Baidu. The reason we also adopt this dataset is that it is currently the largest dataset available for relation extraction. There are 194,747 sentences in the train set and 21,639 in the validate set. We manually labeled 1,150 sentences from the test set with 2,765 annotated triples, which we refer to as SKE21. No preprocessing for this dataset is needed. We provide this data for future research131313download url..
NYT10-HRL | NYT11-HRL | NYT21 | SKE21 | |||||||||
Prec. | Rec. | F1 | Prec. | Rec. | F1 | Prec. | Rec. | F1 | Prec. | Rec. | F1 | |
KB Match | 38.10 | 32.38 | 34.97 | 47.92 | 31.08 | 37.7 | 47.92 | 29.56 | 36.57 | 69.12 | 28.1 | 39.96 |
MultiR Hoffmann et al. (2011) | - | - | - | 32.8 | 30.6 | 31.7 | - | - | - | - | - | - |
SPTree Miwa and Bansal (2016) | 49.2 | 55.7 | 52.2 | 52.2 | 54.1 | 53.1 | - | - | - | - | - | - |
NovelTagging Zheng et al. (2017) | 59.3 | 38.1 | 46.4 | 46.9 | 48.9 | 47.9 | - | - | - | - | - | - |
CoType Ren et al. (2017) | - | - | - | 48.6 | 38.6 | 43.0 | - | - | - | - | - | - |
CopyR Zeng et al. (2018b) | 56.9 | 45.2 | 50.4 | 34.7 | 53.4 | 42.1 | - | - | - | - | - | - |
HRL Takanobu et al. (2019) | 71.4 | 58.6 | 64.4 | 53.8 | 53.8 | 53.8 | - | - | - | - | - | - |
TPLinker Wang et al. (2020)* | 81.19 | 65.41 | 72.45 | 56.2 | 55.14 | 55.67 | 59.78 | 55.78 | 57.71 | - | - | - |
CasRel Wei et al. (2020)* | 77.7 | 68.8 | 73.0 | 50.1 | 58.4 | 53.9 | 58.64 | 56.62 | 57.61 | - | - | - |
ReRe - LSTM | 56.71 | 42.00 | 48.26 | 56.46 | 35.4 | 43.52 | 62.06 | 37.01 | 46.37 | - | - | - |
ReRe | 75.45 | 72.50 | 73.95 | 53.12 | 59.59 | 56.23 | 57.69 | 61.69 | 59.62 | - | - | - |
TPLinker Wang et al. (2020)*(exact) | 80.34 | 65.11 | 71.93 | 55.43 | 55.12 | 55.28 | 58.96 | 55.78 | 57.33 | 83.86 | 84.77 | 84.32 |
CasRel Wei et al. (2020)*(exact) | 75.12 | 65.72 | 70.11 | 47.88 | 55.13 | 51.25 | 55.06 | 54.49 | 54.78 | 86.94 | 85.96 | 86.45 |
ReRe (exact) | 74.90 | 71.97 | 73.4 | 52.40 | 58.91 | 55.47 | 56.97 | 60.93 | 58.88 | 90.44 | 84.20 | 87.21 |
5.2 Compared Methods and Metrics
We evaluate our model by comparing with several models on the same datasets, which are SOTA graphical model MultiR Hoffmann et al. (2011), joint models SPTree Miwa and Bansal (2016) and NovelTagging Zheng et al. (2017), recent strong SOTA models CopyR Zeng et al. (2018b), HRL Takanobu et al. (2019), CasRel Wei et al. (2020), TPLinker Wang et al. (2020). We also provide the result of automatically aligning Wikidata/CN-KBpedia with the corpus, namely Match, as a baseline. To note, we only keep the intersected relations, otherwise it will result in low precision due to the false negative in the original dataset. We report standard micro Precision (Prec.), Recall (Rec.) and F1 score for all the experiments. Following the previous works Takanobu et al. (2019); Wei et al. (2020), we adopt partial match on these data sets for fair comparison. We also provide the results of exact match results of the methods we implemented, and only exact match on SKE2019.
5.3 Overall Comparison
We show the overall comparison result in Table 3. First, we observe that ReRe consistently outperforms all the compared models. We find an interesting result that by purely aligning the database with the corpus, it already achieves surprisingly good overall result (surpassing MultiR) and relatively high precision (comparable to CoType in NYT11-HRL). However, the recall is quite low, which accords with our discussion in Section 2.1 that distant supervision leads to many false negatives. We also provide an ablation result where BERT is replaced with a bidirectional LSTM encoder Graves et al. (2013) with randomly initialized weights. From the results we discover that even without BERT, our framework achieves competitive results against the previous approaches such as CoType and CopyR. This further prove the effectiveness of our ReRe framework.


5.4 How Robust is ReRe against False Negatives?
To further study how our model behaves when training data includes different quantity of false negatives, we conduct experiments on synthetic datasets. We construct five new training data by randomly removing triples with probability of 0.1, 0.3 and 0.5, simulating the situation of different FN rates. We show the precision-recall curves of our method in comparison with CasRel Wei et al. (2020), the best performing competitor, in Figure 3. 1) The overall performance of ReRe is superior to competitor models even when trained on a dataset with a 0.5 FN rate. 2) We show that the intervals of ReRe between lines are smaller than CasRel, indicating that the performance decline under different FN rates of ReRe is smaller. 3) The straight line before curves of our model means that there is no data point at the places where recall is very low. This means that our model is insensitive with the decision boundary and thus more robust.
6 Conclusion
In this paper, we revisit the negative data in relation extraction task. We first show that the false negative rate is largely underestimated by previous researches. We then systematically compare three commonly adopted paradigms and prove that our paradigm suffers less from the overwhelming negative labels. Based on this advantage, we propose ReRe, a pipelined framework that first detect relations at sentence level and then extract entities for each specific relation and provide a multi-label PU learning loss to recover false negatives. Empirical results show that ReRe consistently outperforms the existing state-of-the-arts by a considerable gap, even when learned with large false negative rates.
References
- Bekker and Davis (2018) Jessa Bekker and Jesse Davis. 2018. Estimating the class prior in positive and unlabeled data through decision tree induction. In Proceedings of AAAI.
- Bollacker et al. (2008) Kurt Bollacker, Colin Evans, Praveen Paritosh, Tim Sturge, and Jamie Taylor. 2008. Freebase: a collaboratively created graph database for structuring human knowledge. In Proceedings of SIGMOD.
- Chan and Roth (2011) Yee Seng Chan and Dan Roth. 2011. Exploiting syntactico-semantic structures for relation extraction. In Proceedings of ACL, pages 551–560.
- Chen (2018) Danqi Chen. 2018. Neural Reading Comprehension and Beyond. Ph.D. thesis, Stanford University.
- Chowdhury and Lavelli (2012) Md. Faisal Mahbub Chowdhury and A. Lavelli. 2012. Impact of less skewed distributions on efficiency and effectiveness of biomedical relation extraction. In Proceedings of COLING.
- Devlin et al. (2019) J. Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of NAACL-HLT.
- Feng et al. (2018) Jun Feng, Minlie Huang, Li Zhao, Yang Yang, and Xiaoyan Zhu. 2018. Reinforcement learning for relation classification from noisy data. In Proceedings of AAAI, volume 32.
- Gardent et al. (2017) Claire Gardent, Anastasia Shimorina, Shashi Narayan, and Laura Perez-Beltrachini. 2017. Creating training corpora for NLG micro-planners. In Proceedings of ACL.
- Gormley et al. (2015) Matthew R Gormley, Mo Yu, and Mark Dredze. 2015. Improved relation extraction with feature-rich compositional embedding models. In Proceedings of ACL, pages 1774–1784.
- Graves et al. (2013) Alex Graves, Abdel-rahman Mohamed, and Geoffrey Hinton. 2013. Speech recognition with deep recurrent neural networks. In 2013 IEEE international conference on acoustics, speech and signal processing, pages 6645–6649.
- Gupta et al. (2016) Pankaj Gupta, Hinrich Schütze, and Bernt Andrassy. 2016. Table filling multi-task recurrent neural network for joint entity and relation extraction. In Proceedings of COLING, pages 2537–2547.
- Hoffmann et al. (2011) R. Hoffmann, Congle Zhang, Xiao Ling, Luke Zettlemoyer, and Daniel S. Weld. 2011. Knowledge-based weak supervision for information extraction of overlapping relations. In Proceedings of ACL.
- Jia et al. (2019) Wei Jia, Dai Dai, Xinyan Xiao, and Hua Wu. 2019. ARNOR: Attention regularization based noise reduction for distant supervision relation classification. In Proceedings of ACL, pages 1399–1408.
- Katiyar and Cardie (2017) Arzoo Katiyar and Claire Cardie. 2017. Going out on a limb: Joint extraction of entity mentions and relations without dependency trees. In Proceedings of ACL, pages 917–928.
- Li and Ji (2014) Q. Li and Heng Ji. 2014. Incremental joint extraction of entity mentions and relations. In Proceedings of ACL.
- Li et al. (2020) Xiaoya Li, Xiaofei Sun, Yuxian Meng, Junjun Liang, F. Wu, and J. Li. 2020. Dice loss for data-imbalanced nlp tasks. In Proceedings of ACL.
- Li et al. (2019) Xiaoya Li, Fan Yin, Zijun Sun, Xiayu Li, Arianna Yuan, Duo Chai, Mingxin Zhou, and J. Li. 2019. Entity-relation extraction as multi-turn question answering. In Proceedings of ACL.
- Lin et al. (2018) Hongyu Lin, Yaojie Lu, Xianpei Han, and Le Sun. 2018. Adaptive scaling for sparse detection in information extraction. In Proceedings of ACL, pages 1033–1043.
- Lin et al. (2016) Yankai Lin, Shiqi Shen, Zhiyuan Liu, Huanbo Luan, and Maosong Sun. 2016. Neural relation extraction with selective attention over instances. In Proceedings of ACL, pages 2124–2133.
- Liu et al. (2019) Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692.
- Min et al. (2013) Bonan Min, Ralph Grishman, Li Wan, Chang Wang, and David Gondek. 2013. Distant supervision for relation extraction with an incomplete knowledge base. In Proceedings of HLT-NAACL.
- Mintz et al. (2009) Mike Mintz, Steven Bills, Rion Snow, and Dan Jurafsky. 2009. Distant supervision for relation extraction without labeled data. In Proceedings of ACL.
- Miwa and Bansal (2016) Makoto Miwa and Mohit Bansal. 2016. End-to-end relation extraction using lstms on sequences and tree structures. In Proceedings of ACL, pages 1105–1116.
- Miwa and Sasaki (2014) Makoto Miwa and Yutaka Sasaki. 2014. Modeling joint entity and relation extraction with table representation. In Proceedings of EMNLP, pages 1858–1869.
- du Plessis et al. (2015) Marthinus Christoffel du Plessis, Gang Niu, and Masashi Sugiyama. 2015. Class-prior estimation for learning from positive and unlabeled data. Machine Learning, 106:463–492.
- Ren et al. (2017) Xiang Ren, Zeqiu Wu, Wenqi He, Meng Qu, Clare R Voss, Heng Ji, Tarek F Abdelzaher, and Jiawei Han. 2017. Cotype: Joint extraction of typed entities and relations with knowledge bases. In Proceedings of WWW, pages 1015–1024.
- Riedel et al. (2010) S. Riedel, Limin Yao, and A. McCallum. 2010. Modeling relations and their mentions without labeled text. In Proceedings of ECML/PKDD.
- Roller et al. (2015) Roland Roller, Eneko Agirre, Aitor Soroa, and Mark Stevenson. 2015. Improving distant supervision using inference learning. In Proceedings of ACL, pages 273–278.
- dos Santos et al. (2015) Cicero dos Santos, Bing Xiang, and Bowen Zhou. 2015. Classifying relations by ranking with convolutional neural networks. In Proceedings of ACL, pages 626–634.
- Takanobu et al. (2019) Ryuichi Takanobu, Tianyang Zhang, Jiexi Liu, and Minlie Huang. 2019. A hierarchical framework for relation extraction with reinforcement learning. In Proceedings of AAAI, volume 33, pages 7072–7079.
- Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Proceedings of NeuroIPS, pages 6000–6010.
- Vrandecic and Krötzsch (2014) Denny Vrandecic and M. Krötzsch. 2014. Wikidata: a free collaborative knowledgebase. Communications of the ACM, 57:78–85.
- Wang and Jiang (2017) Shuohang Wang and Jing Jiang. 2017. Machine comprehension using match-lstm and answer pointer. In Proceedings of ICLR.
- Wang et al. (2020) Yucheng Wang, Bowen Yu, Yueyang Zhang, Tingwen Liu, Hongsong Zhu, and Limin Sun. 2020. TPLinker: Single-stage joint extraction of entities and relations through token pair linking. In Proceedings of COLING, pages 1572–1582.
- Wei et al. (2020) Zhepei Wei, Jianlin Su, Yue Wang, Y. Tian, and Yi Chang. 2020. A novel cascade binary tagging framework for relational triple extraction. In Proceedings of ACL.
- Xie et al. (2020) Chenhao Xie, Qiao Cheng, Jiaqing Liang, Lihan Chen, and Y. Xiao. 2020. Collective loss function for positive and unlabeled learning. ArXiv, abs/2005.03228.
- Xu et al. (2017) Bo Xu, Yong Xu, Jiaqing Liang, Chenhao Xie, Bin Liang, Wanyun Cui, and Y. Xiao. 2017. CN-DBpedia: A never-ending chinese knowledge extraction system. In Proceedings of IEA/AIE.
- Xu et al. (2013) Wei Xu, Raphael Hoffmann, Le Zhao, and Ralph Grishman. 2013. Filling knowledge base gaps for distant supervision of relation extraction. In Proceedings of ACL, pages 665–670.
- Yu et al. (2020) Bowen Yu, Zhenyu Zhang, Xiaobo Shu, Tingwen Liu, Yubin Wang, Bin Wang, and Sujian Li. 2020. Joint extraction of entities and relations based on a novel decomposition strategy. In Proceedings of ECAI.
- Yu and Lam (2010) Xiaofeng Yu and Wai Lam. 2010. Jointly identifying entities and extracting relations in encyclopedia text via a graphical model approach. In Proceedings of COLING, pages 1399–1407.
- Zeng et al. (2015) Daojian Zeng, Kang Liu, Yubo Chen, and Jun Zhao. 2015. Distant supervision for relation extraction via piecewise convolutional neural networks. In Proceedings of EMNLP.
- Zeng et al. (2014) Daojian Zeng, Kang Liu, Siwei Lai, Guangyou Zhou, Jun Zhao, et al. 2014. Relation classification via convolutional deep neural network. In Proceedings of COLING, pages 2335–2344.
- Zeng et al. (2018a) Xiangrong Zeng, Shizhu He, Kang Liu, and Jun Zhao. 2018a. Large scaled relation extraction with reinforcement learning. In Proceedings of AAAI, volume 32.
- Zeng et al. (2018b) Xiangrong Zeng, Daojian Zeng, Shizhu He, Kang Liu, and Jun Zhao. 2018b. Extracting relational facts by an end-to-end neural model with copy mechanism. In Proceedings of ACL.
- Zhao et al. (2020) Tianyang Zhao, Zhao Yan, Y. Cao, and Zhoujun Li. 2020. Asking effective and diverse questions: A machine reading comprehension based framework for joint entity-relation extraction. In Proceedings of IJCAI.
- Zheng et al. (2017) Suncong Zheng, Feng Wang, Hongyun Bao, Yuexing Hao, Peng Zhou, and Bo Xu. 2017. Joint extraction of entities and relations based on a novel tagging scheme. In Proceedings of ACL, pages 1227–1236.