Evaluating Factuality in Generation with Dependency-level Entailment
Abstract
Despite significant progress in text generation models, a serious limitation is their tendency to produce text that is factually inconsistent with information in the input. Recent work has studied whether textual entailment systems can be used to identify factual errors; however, these sentence-level entailment models are trained to solve a different problem than generation filtering and they do not localize which part of a generation is non-factual. In this paper, we propose a new formulation of entailment that decomposes it at the level of dependency arcs. Rather than focusing on aggregate decisions, we instead ask whether the semantic relationship manifested by individual dependency arcs in the generated output is supported by the input. Human judgments on this task are difficult to obtain; we therefore propose a method to automatically create data based on existing entailment or paraphrase corpora. Experiments show that our dependency arc entailment model trained on this data can identify factual inconsistencies in paraphrasing and summarization better than sentence-level methods or those based on question generation, while additionally localizing the erroneous parts of the generation.111Data and code available at https://github.com/tagoyal/dae-factuality
1 Introduction
The rise of pre-trained language models Devlin et al. (2019); Radford et al. (2019) has led to strong text generation models for applications including summarization Dong et al. (2019); Lewis et al. (2020), paraphrasing Goyal and Durrett (2020); Shen et al. (2020), story generation Mao et al. (2019), and data augmentation Yu et al. (2018); Zhang and Bansal (2019). However, while these models generate fluent and grammatical text, they are prone to making factual errors that contradict the input text Cao et al. (2018). Automatic metrics used to evaluate text generation, such as ROUGE and BERTScore Zhang et al. (2020), are not correlated with the factual consistency or faithfulness of the generated text Falke et al. (2019); Kryściński et al. (2019). To address this, recent work has studied the use of textual entailment models to rank and filter non-factual generations Falke et al. (2019); Maynez et al. (2020). However, these models suffer from issues such as dataset biases Gururangan et al. (2018); Zhou and Bansal (2020) and a mismatch between the training data (entailment) and the test data (model generations).

In this paper, we propose to decompose entailment decisions in a sentence to evaluate the faithfulness of generated text in a more fine-grained way. Rather than making a sentence-level entailment decision, we instead evaluate the entailment of dependency arcs of the generated sentence, as illustrated in Figure 1. This approach views dependency arcs as semantic units that can be interpreted in isolation. Each arc is therefore judged independently based on whether the relation it implies is entailed by the source sentence. This is helpful in localizing generation errors and consequently providing more interpretable model decisions.
Decomposing the factuality evaluation over components of structured representations can also be extended to other formalisms like AMR Banarescu et al. (2013), UDS White et al. (2016), and more. The chief advantage of dependency parsing over these is that pre-existing tools for dependency parsing report very high performance. Another line of work focuses on question answering-based semantic representations FitzGerald et al. (2018); Michael et al. (2018) or generating freeform questions to capture factuality Wang et al. (2020); Durmus et al. (2020). However, these systems require a separate question generation step at inference time, so generation is baked into their formalisms in a heavyweight way whereas we only require dependencies. A final advantage of our approach is that dependency arcs can be produced in an online fashion during inference, and hence, factuality can be enforced incrementally.
We evaluate our proposed dependency arc entailment approach in both summarization and paraphrase settings. In both settings, we show that we can automatically derive labels from actual generation data rather than rely on human annotation of dependency arc entailment, which is challenging to collect at scale. Nevertheless, our results show that our system’s performance on factuality classification surpasses both sentence-level entailment and question generation and answering models. Our derived labels from actual generation data provide much better task-specific supervision compared to general entailment datasets. Finally, we demonstrate that predicted entailment scores for individual dependency arcs are meaningful and can be leveraged to understand and localize errors in system generations.
2 Dependency Arc Entailment (DAE)
Defining arc entailment
Our notion of entailment starts by assuming a rough correspondence between predicates and arguments in two sentences. In natural language inference (NLI) annotation efforts, this has taken the form of anchoring judgments in an underlying imagined scene Bowman et al. (2015). We make a similar assumption, that events and actors in the source and target sentences are in correspondence unless there is direct evidence to the contrary. For instance, in Figure 1, the military coup in the target sentence and its corresponding amod(coupmilitary) arc should be evaluated with respect to the military takeover in the source, giving coreference of the two the benefit of the doubt here.
With this assumption, we say that a dependency arc in the target sentence is entailed by the source if the semantic relationship it implies between its head and child is entailed by the source sentence. There is precedent for such a syntax-semantics correspondence: certain formalisms like meaning-text theory Mel’čuk (1988) have historically made this mapping more or less explicit. Consider the first hypothesis in Figure 1. Many of the arcs here either contain information analogous to that in semantic roles, or they specify nominal modifiers capturing important entity properties.222We use enhanced dependencies in our experiments; modifiers with prepositions, augmented conjuncts provide a more useful semantic representation. In our implementation, we exclude certain arc types which are not strongly tied to semantics, such as arcs involving punctuation; see the Appendix for details. Note that our method does not support commenting on arcs of the input that do not exist in the output; we discuss this later in Section 7.2.

In some ways, our view of entailment is equivalent with the entailment of NLI settings Bowman et al. (2015); Williams et al. (2018): if a hypothesis is entailed under the NLI definition, then all dependency arcs of the hypothesis must be entailed by our DAE definition. However, in our formulation, arc entailment is a 2-class classification task with labels {entailed, non-entailed}. This means that arcs that would be neutral or contradiction in the generic entailment formulation are considered non-entailed in our scenario.
Annotating arc entailment
To model this formulation, we require entailment annotations at the dependency arc level. However, there are several challenges associated with human annotation of arc entailment data. (1) Entailment is not truly a binary decision and is inherently subjective Pavlick and Kwiatkowski (2019). (2) Entailment of an arc may be fundamentally unknowable or undefined if, for example, too much of the context has changed for such a judgment to make sense. (3) Annotators would need to understand the meaning of dependency labels and to be able to isolate the semantics of these individual arcs in sentences.
P: Australian Prime Minister, John Howard, has made an unannounced visit to Iraq, according to the office of Iraqi transitional Prime Minister Ibrahim al-Jaafari. |
H: Howard is a political representative of Australia. |
Therefore, in this work, we take another approach, which is to automatically label data from existing sources and outputs of generation models, which lets us collect large-scale data in a variety of domains. Specifically, we use paraphrase data to construct our dataset. Note however, that there is a fundamental difference between paraphrase pairs, which should be entailed in both directions, and past NLI data, which is forward-entailed by definition. For instance, the premise and hypothesis in Table 1 would classically be judged as entailed because political representative is a hypernym of prime minister, but the hypothesis is not a paraphrase of (even part of) the premise.
As a result, our automatically-derived dataset captures a more restricted notion of entailment, primarily consisting of entailment relations that are symmetric in nature: arcs in the target sentence entailed by a source sentence also entail some part of the source. However, this is actually closer to what is acceptable for generation models to produce in tasks such as summarization, and the dataset collected in such a manner is useful for downstream tasks, as we show in Section 6. Moreover, because our training and evaluation data will typically come from closely related sentences, we can sidestep the cases where judgments in our formalism become most difficult to define.
3 Model
Let be the input context, be a hypothesis produced by a generation model , and be the set of arcs in the dependency parse of . We want to predict the entailment decision for each arc with respect to the input , i.e. .
The overall model architecture of this dependency arc entailment model is outlined in Figure 2. First, we concatenate the input and the hypothesis. We use a pre-trained encoder model to obtain contextual representations for each token in the concatenated sequence. From these token level representations, we derive a representation for each dependency arc :
as shown in the inset in the figure. Here, , are the token indices corresponding to the head word and the child word of dependency arc , and is their corresponding dependency label, which is also embedded with (non-contextually).
Next, these arc representations are passed through a linear layer, followed by a softmax layer to obtain entailment label probabilities corresponding to each arc: .
This DAE network is trained using standard binary cross entropy loss and requires supervision on the arc entailment labels {entailed, non-entailed} for the dependency arcs. Examples do not need to be fully labeled; training can use partial sets of annotations of arcs in . However, while using the DAE model in downstream tasks such as hypotheses reranking, entailment decisions for all arcs in the candidate hypothesis are required.
Sentence-level factuality from dependency-level judgments
We want to evaluate the factual consistency of each hypothesis with respect to input , i.e. . This is computed by combining arc-level entailment scores over the dependency arcs set of the generated hypothesis:
We use the sentence-level score to rerank the generated hypotheses in Section 6.333According to DAE definition, an output is non-factual if any of its arcs is non-entailed. However, min-pooling was very unstable, so we instead use mean-pooling in our experiments.
4 Automatic Dataset Creation
We now describe our method for automatically collecting dependency-level DAE annotations from paraphrase or entailment corpora, avoiding manual annotation. In this creation process, we want data featuring a range of paraphrasing phenomenon, such as passivization, clausal reordering, synonym replacement, and more. Furthermore, we want a natural distribution of errors produced by generation models, such as wrong subject or objects for a verb or hallucination of new content.
We represent a single example in our dataset as a tuple . Here, is the input, is the hypothesis, denotes a single dependency arc in the hypothesis, and refers to the gold entailment label for that arc. To construct data of this form, we assume access to a paraphrase dataset , containing pairs of input sentences and their corresponding gold paraphrases.444The paraphrase corpora we use in this work may come from automatic methods like backtranslation; however, we still assume that these are reliable gold-standard paraphrases. Additionally, we employ a paraphrase generation model , which can output candidate paraphrases given an input . These noisy paraphrases will be used to derive realistic examples of generation errors to contrast with gold paraphrases, using the following techniques.
Positive labels from gold paraphrases
Given a ground truth paraphrase, we assume that every arc in the target side of the paraphrase is entailed by the source side . This is in line with our definition of arc entailment in Section 2 and allows us to propagate sentence-level paraphrase judgements to arc-level entailment judgements. Because paraphrase datasets feature diverse linguistic phenomena, this approach leads to a range of positive examples. However, as described in Section 2, it is less likely to include arcs which are forward-entailed only (e.g., Table 1).
Auto-derived labels from model generations
To find negative examples for entailment, we leverage the output generations of an automatic paraphrase model . These generations will include unseen arcs, which may be positively or negatively entailed.
Our key assumption here is that the outputs at the top of the beam are more likely to be factually correct, whereas outputs at the bottom of the beam are of lower quality and more prone to having factual inconsistencies. We assume that new arcs introduced in bad model generations (i.e., bottom-most generations of the beam) are not entailed by the input.
We can then noisily label the generated paraphrases with a mix of positive (entailed) and negative (non-entailed) labels. We first construct a set of entailed dependency arcs: this is a set containing all dependency arcs of the input and the gold paraphrase, i.e., . Next, we annotate the dependency arcs of the bottom-most generations of the beam, say , in the following way:
The middle case here leaves arcs that are in but not or as unannotated. Such arcs are possibly factual under the model, coming from a high-quality generated output, but we do not have enough confidence to assign them a label. During model training, such unannotated arcs are ignored.
Finally, we also include the positive arcs from the 1-best hypothesis in our DAE data: for arcs . This provides another source of hypothesis sentences with a slightly different distribution during model training.
5 Intrinsic Evaluation of DAE
Our experimental evaluation focuses on the following questions: (1) Does the automatic data collection result in a high quality training dataset? (2) Is the DAE model we train a good classifier? (3) Does DAE allow us to filter model generations and improve reliability of a generation system?
We construct our DAE training dataset using the methodology defined in Section 4. For this, we leverage the paraphrase pair dataset Paranmt-50m Wieting and Gimpel (2018) as the base dataset . We use the transformer-based encoder-decoder model for paraphrase generation from Goyal and Durrett (2020) as . We use the paraphrase model to generate outputs for 20k sentence pairs from . We use the Stanford CoreNLP library Manning et al. (2014) to extract enhanced dependencies from the outputs sentences. Then, using the strategy outlined in Section 4, we generate 100k training samples (sentence pairs), 3k dev samples and 3k test samples. From this dataset, we derive 520k training, 14k dev, and 22k dependency level annotations, which we evaluate on in Section 5.2. The entailed to not-entailed ratio is roughly 70-30 in this dataset.
5.1 Dataset Quality
Before evaluating our modeling approach, we first evaluate whether the arc annotations in the training data follow the theoretical definition of DAE, outlined in Section 2. Figure 3 showcases examples from the dev set, corresponding to the same input example. We show positive entailed arcs (in green), negative non-entailed arcs (in red), and one unlabeled arc (in gray). Here, we can see that the gold paraphrase is important as it provides examples of valid synonym replacements, as well as other rephrasing of the input sentence. For negative examples, the examples from the bottom of the beam do indeed contain bad output and non-entailed arcs.

Agreement with human labels
Next, we want to evaluate the quality of the auto-derived dataset by measuring its agreement with human annotations. For this, we manually annotated the dependency arc entailment labels for 100 sentence pairs from the dev set (consisting of 20 gold paraphrases and 80 generated paraphrases), according to our theoretical definition. We compared these manual annotations (gold) with the auto-derived annotations for this set, and observed that the two annotations agreed 82% of the time. This indicates that the automatic annotation strategy from Section 4 results in a high quality dataset. Further investigation into the disagreements between the manual and automatic labels revealed that false negatives included paraphrasing phenomena like synonym replacement, anaphora resolution during reordering, etc. We describe how to produce additional data to handle some of these cases later. On the other hand, false positives mainly consisted of exact arc matches in incorrect contexts.
5.2 Intrinsic Evaluation: DAE Classification
Model | Accuracy | F1 |
---|---|---|
Majority label | 72.7 | 84.2 |
Lexical-match | 74.2 | 78.1 |
Bert | 86.9 | 91.0 |
Electra | 88.4 | 92.1 |
Next, we intrinsically evaluate the performance of the dependency arc entailment model, outlined in Section 3, on held-out data from our automatic labeling method. We test the performance of two pre-trained models: Bert (bert-base-uncased, 110M parameters) Devlin et al. (2019) and Electra (electra-base-discriminator, 110M parameters) Clark et al. (2020). We compare these models against a majority label baseline (entailment) and an lexical-match baseline that predicts if the arc (head, child and label) in the output constitute a dependency arc in the input as well, and non-entailed otherwise.
The performance of the different models is outlined in Table 2. Our pre-trained transformer models perform substantially better than the baselines, with Bert achieving accuracy, and Electra with . These models also outperform the lexical-match baseline, showing that the DAE models learn to do more than simple dependency arc matching. Henceforth, we use the best performing Electra model in all our experiments.
5.3 Other Data Generation Methods
Besides the data generation procedure we proposed, there are other ways to synthetically generate noisy annotations for premise-hypothesis pairs Zhang et al. (2019). We investigate these from two perspectives: first, does our data generation process cover these phenomena well, and second, can these additional sources of data prove useful?
First, we explore a word-swapping technique similar to Zhang et al. (2019). Given a premise , we form a hypothesis by randomly swapping tokens that share a common part-of-speech tag to introduce errors. The intersection of the arcs in and the modified sentence are annotated as positive arcs (), whereas the newly created or changed arcs are annotated as negative ().
Our synonym data is noisily constructed in the same manner as the gold paraphrases, but targets synonym replacements specifically. We extract pairs from that generally maintain similar sentence structure between the two sentences,555We follow prior work Goyal and Durrett (2020) and calculate structure similarity by aligning words in the input and target using GloVe Pennington et al. (2014) and computing the average displacement of each word. but with small lexical changes like synonym replacement. We assign a positive entailment label to all arcs: .
To construct data with hallucinations, we modify an input sentence , which we take as the hypothesis by removing a randomly sampled span of contiguous tokens to derive a premise sentence . Then, the following DAE model annotations are derived: . Additionally, for each input sentence , we extract another sentence with the highest 1-gram overlap in the dataset. From this we derive, .
Test set | ||||
---|---|---|---|---|
Model Training Source | WS | AD | S | H |
Word-swapping (WS) | 98.5 | 71.6 | 29.6 | 80.0 |
Auto-derived (AD) | 90.2 | 88.4 | 82.9 | 74.8 |
+ synonyms (S) | 90.5 | 88.0 | 96.0 | 73.9 |
+ hallucinations (H) | 92.4 | 87.8 | 96.9 | 97.6 |
Table 3 shows a comparison of word-swapping with our method (AD), and variants of our method augmented with synonyms and hallucinations. Note that the model trained on word swapped data performs well on a similarly constructed held-out set, but not on the test data for synonym data and auto-derived data. This indicates that artificially constructed data with rule based error introduction does not cover the space of generation possibilities. On the other hand, the model trained on our auto-derived dataset performs well across both artificial and actual generation data, thereby covering a larger range of entailment possibilities. Additional augmentation of synonym- and hallucination-specific data improves the performance further on the respective test sets while retaining the performance on generic entailment data. Henceforth, we use the (AD + S) model for our experiments.
6 Extrinsic Evaluation: Filtering Bad Generations
Moving beyond the dependency-level inference task, we now want to evaluate the sentence-level performance of our model formulation. Namely, can it usefully reject erroneous generations produced by models for summarization (Section 6.1) and paraphrasing (Section 6.2)?
6.1 Summary Ranking
We perform our evaluation on an abstractive summarization test dataset introduced in Falke et al. (2019) and used in other previous work. It contains test samples, each containing an input source sentence from CNN/DM and two summary sentences covering the same content generated using the model from Chen and Bansal (2018). One of these summary sentences is factually correct and the other is factually incorrect. The evaluation protocol measures how often the correct summary is scored higher than the incorrect summary for each candidate scoring technique. We compare against the following baselines:
-
1.
NLI models: Following Falke et al. (2019), we use entailment predictions of NLI models to rerank candidates. We compare the performance of pretrained encoders (Bert, RoBerta and Electra) fine-tuned on the MNLI dataset Williams et al. (2018).666We fine-tune the Bert and Electra models ourselves (details in the Appendix), improving on the results from Falke et al. (2019). We use the fine-tuned RoBERTa model released by AllenNLP (https://allennlp.org/).
-
2.
Question Generation and Answering: Wang et al. (2020) propose an automatic evaluation metric QAGS that scores each summary by first generating questions pertaining to the summary content, and then comparing the answers to those questions in both the source sentence and the generated summary.
-
3.
Rule-based: We score each summary sentence as the fraction of dependency arcs in the output that are common with the input sentence. In case both the correct and the incorrect sentence get the same score, we break ties randomly.
Model | Reranking Acc. |
---|---|
BERT (MNLI) | 72.3 |
RoBERTa (MNLI) | 78.3 |
Electra (MNLI) | 82.0 |
QAGS | 72.1 |
Rule-based dependency | 74.8 |
DAE (ours) | 83.6 |
Human | 83.9 |
Table 4 outlines the performance of the different models. The results show that the dependency arc entailment model outperforms the sentence-level NLI models and also the question generation and answering formulation (QAGS). Furthermore, the performance of our DAE model is close to the human performance reported in Falke et al. (2019). Interestingly, the rule-based dependency model also outperforms certain NLI models and QAGS, indicating that these more complex models may fail to capture straightforward lexical relations.
During our experimentation, we observed large variance in the performance of the NLI models at the reranking task with respect to their performance at the intrinsic entailment task. To illustrate this, in Figure 4, we plot the summarization reranking performance of the two model against the intrinsic task performance at different stages of the training. For DAE, the intrinsic task performance is reported by the dependency-level entailment classification accuracy, and for the MNLI model, we report the classification accuracy on the sentence-level MNLI entailment task. The graph shows a high variance in the summary reranking performance, with stable increase in the MNLI intrinsic task performance at different time steps.777Note that the best performance of the MNLI model on summary reranking is better than the best performance of the DAE model; however, it did not coincide with the task-level best performance for our particular hyperparameter choice. This indicates that the general entailment task solves a fundamentally different problem than factuality. By contrast, the DAE model performance on the summarization reranking task is more stable.

6.2 Paraphrase Ranking
Next, we evaluate the DAE model in the paraphrasing setting. To do this, first, we create a test set, similar to the summarization test set from Falke et al. (2019). We use the transformer based seq2seq model Goyal and Durrett (2020) to obtain candidate paraphrases for input sentences from the ParaNMT-50m dataset. We manually assign a label factual, not factual to each input, candidate pair. Then for each input sentence, we randomly selected one correct and one incorrect paraphrase. This sentence triplet is used for our reranking experiments.
Model | Reranking Acc |
---|---|
MNLI (Electra) | 79.0 |
DAE (Electra) | 79.0 |
Table 5 compares the performance of the MNLI-based model and the DAE models. Here, both are Electra based models; these are shown to be the best performing models in the previous sections. The results show that in this setting, the MNLI model and the DAE model perform similarly. Closer inspection of this data revealed that our model is biased towards predicting the label entailment for arcs that are common with the input, possibly because we are evaluating the same generation model that was used to produce our arc entailment data, and our model is therefore biased towards predicting non-entailed for arcs that are not present in the input. This poses a somewhat adversarial setting for our DAE model.
7 Analysis

7.1 Dependency- vs. sentence-level modeling
Although our DAE model has shown strong performance, we have not yet performed a direct apples-to-apples comparison of DAE versus a sentence-level model when trained on the same sentences.
For MNLI
We construct DAE data from the sentence-level entailment data as follows. First, we extract 10k examples from the MNLI data which have the label entailment. This is considered as the source data . We use a paraphrase model (transformer seq2seq Goyal and Durrett (2020)) and the technique outlined in Section 4 to extract auto-derived labels from . This gives us 42k training examples for training the DAE model. We compare this against an MNLI model trained on the original sentence-level entailment task with the same number of examples (42k).
For ParaNmt
For this dataset, we do not have negative () annotations at the sentence-level. We derive these from our DAE training data as follows: we consider all pairs of sentences in the original dataset () as positive sentences (), in addition to any pair of the form (). We treat the three generated sentences at the bottom of the beam as negative sentences, meaning that the model is trained to distinguish gold paraphrases from model generations.
Model | ParaNMT | MNLI | ||
---|---|---|---|---|
Summ | Para | Summ | Para | |
sent-level | 73.9 | 58.0 | 68.8 | 64.0 |
dep-level | 83.6 | 79.0 | 78.5 | 79.0 |
Table 6 outlines these results. For the paraphrase dataset, we see that the artificially constructed sentence-level dataset does not yield a good sentence-level discriminator. However, our dependency-level annotations can form an effective training set. The results on MNLI show that our dependency-level formulation performs better than sentence-level when trained on the same amount of data and is therefore more closely related to the factuality task than past entailment formulations.
7.2 Qualitative Evaluation
Error Localization
Since the DAE formulation computes individual entailment scores for all arcs in the dependency tree structure, it is possible to localize errors in the generated summary or paraphrase. We present examples of input sentences, generated text, and arc entailment scores for a few examples in Figure 5. For each input and output pair, we show the individual scores for the dependency arcs in the output sentence. Additionally, we report the MNLI score for the same example.
The illustrative examples show that the DAE model is capable of localizing errors where erroneous subject-object pairs were constructed, even when these are the only errors. These errors are tougher for the MNLI model to catch, which evaluates the whole sentence and is prone to lexical overlap biases Zhou and Bansal (2020). In the third example, from our paraphrase setting, we see that the model is able to recognize synonym replacement as a valid re-writing. However, for the last example, the model cannot perform this same judgement for the variations changes replacement. Although, note that the model scores it higher than a erroneous replacement of the same word (variations latter). This shows that the DAE model is able to rank sentences that incorporate the similar type of re-writing/editing. However, we observed that the model has different error rates for different types of re-writing changes. For example, it is better at identifying text hallucination, or cases where the subject object relation between words change, but has comparatively lesser accuracy over changes such as synonym replacements. Therefore, it may not be ideal for settings where different re-writing types need to be compared.
Limitations
We comment on a few limitations of our approach. First, arcs in our dependency-based formalism are not marked with negation or quantification; these must be handled via the contextualization of the hypothesis sentence rather than in the semantic representation. Second, our method cannot identify arcs that are missing from the input. For instance, consider the following premise: In the morning, he goes jogging and hypothesis: In the morning. Here, the hypothesis does not contain critical information from the source sentence; however, since all the present arcs are entailed by the source, our model would give this a high score.
Furthermore, our model is trained on the ParaNMT-50m dataset, which itself is constructed through a noisy backtranslation process. Therefore, we rely on noisy gold data for constructing our model. We believe that better quality paraphrase pairs would lead to a better quality model.
8 Related Work
Recent work in addressing faithfulness of text generations can be broadly divided into three groups: structured information based, multi-task formulations, and post-processing methods. The first group leverages structured knowledge, like Open IE triples Cao et al. (2018); Goodrich et al. (2019), dependency trees Song et al. (2018), or generated semantic roles Fan et al. (2018) as additional input for generation. However, incorporation of these as additional embeddings in model architectures does not explain how these influence model generations. The second group leverages multi-task formulations and trains the generation model jointly with other factuality-related tasks, such as NLI entailment and question generation Guo et al. (2018). Other work additionally incorporates a reward for generating summaries entailed by the the input Li et al. (2018); Pasunuru and Bansal (2018). Our approach can be used to rank/filter outputs from any generation model in a black-box way, without additional augmentation or retraining.
In post-processing approaches, recent work has explored NLI-based Falke et al. (2019); Maynez et al. (2020) post-generation filtering or ranking of output summaries. Our dependency-level models perform on par with these approaches, while additionally localizing the error in the generations. Other work Durmus et al. (2020); Wang et al. (2020) has looked at using question generation and answering to reveal factual inconsistencies in generated text. However, more work is needed to figure out how to make these approaches reliable and broad coverage, as they primarily focus on specific factors like noun phrases and named entities.
9 Conclusion
In this work, we propose the dependency arc entailment formulation to identify factual errors in generated text in a more fine-grained manner. We show that the proposed formulation outperforms past approaches, while additionally providing an interpretable error analysis.
Acknowledgments
Thanks to Katrin Erk for providing feedback on a draft of this work, as well to the anonymous reviewers for their helpful comments. This work was partially supported by NSF Grant IIS-1814522, a gift from Salesforce Inc, and an equipment grant from NVIDIA. The authors acknowledge the Texas Advanced Computing Center (TACC) at The University of Texas at Austin for providing HPC resources used to conduct this research.
References
- Banarescu et al. (2013) Laura Banarescu, Claire Bonial, Shu Cai, Madalina Georgescu, Kira Griffitt, Ulf Hermjakob, Kevin Knight, Philipp Koehn, Martha Palmer, and Nathan Schneider. 2013. Abstract meaning representation for sembanking. In Proceedings of the 7th linguistic annotation workshop and interoperability with discourse, pages 178–186.
- Bar-Heim et al. (2006) Roy Bar-Heim, Ido Dagan, Bill Dolan, Lisa Ferro, Danilo Giampiccolo, Bernardo Magnini, and Idan Szpektor. 2006. The Second PASCAL Recognizing Textual Entailment Challenge. In Proceedings of the Second PASCAL Challenges Workshop on Recognising Textual Entailment, Venice, Italy, page 49.
- Bowman et al. (2015) Samuel R. Bowman, Gabor Angeli, Christopher Potts, and Christopher D. Manning. 2015. A large annotated corpus for learning natural language inference. In Empirical Methods in Natural Language Processing (EMNLP).
- Cao et al. (2018) Ziqiang Cao, Furu Wei, Wenjie Li, and Sujian Li. 2018. Faithful to the original: Fact aware neural abstractive summarization. In Thirty-Second AAAI Conference on Artificial Intelligence.
- Chen and Bansal (2018) Yen-Chun Chen and Mohit Bansal. 2018. Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers).
- Clark et al. (2020) Kevin Clark, Minh-Thang Luong, Quoc V. Le, and Christopher D. Manning. 2020. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators. In International Conference on Learning Representations.
- Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers).
- Dong et al. (2019) Li Dong, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Jianfeng Gao, Ming Zhou, and Hsiao-Wuen Hon. 2019. Unified language model pre-training for natural language understanding and generation. In Advances in Neural Information Processing Systems, pages 13042–13054.
- Durmus et al. (2020) Esin Durmus, He He, and Mona Diab. 2020. FEQA: A Question Answering Evaluation Framework for Faithfulness Assessment in Abstractive Summarization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.
- Falke et al. (2019) Tobias Falke, Leonardo FR Ribeiro, Prasetya Ajie Utama, Ido Dagan, and Iryna Gurevych. 2019. Ranking generated summaries by correctness: An interesting but challenging application for natural language inference. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 2214–2220.
- Fan et al. (2018) Lisa Fan, Dong Yu, and Lu Wang. 2018. Robust Neural Abstractive Summarization Systems and Evaluation against Adversarial Information. In Workshop on Interpretability and Robustness in Audio, Speech, and Language (IRASL). Neural Information Processing Systems.
- FitzGerald et al. (2018) Nicholas FitzGerald, Julian Michael, Luheng He, and Luke Zettlemoyer. 2018. Large-Scale QA-SRL Parsing. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2051–2060.
- Goodrich et al. (2019) Ben Goodrich, Vinay Rao, Peter J Liu, and Mohammad Saleh. 2019. Assessing the factual accuracy of generated text. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 166–175.
- Goyal and Durrett (2020) Tanya Goyal and Greg Durrett. 2020. Neural Syntactic Preordering for Controlled Paraphrase Generation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.
- Guo et al. (2018) Han Guo, Ramakanth Pasunuru, and Mohit Bansal. 2018. Soft Layer-Specific Multi-Task Summarization with Entailment and Question Generation. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 687–697.
- Gururangan et al. (2018) Suchin Gururangan, Swabha Swayamdipta, Omer Levy, Roy Schwartz, Samuel Bowman, and Noah A Smith. 2018. Annotation Artifacts in Natural Language Inference Data. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers).
- Kryściński et al. (2019) Wojciech Kryściński, Nitish Shirish Keskar, Bryan McCann, Caiming Xiong, and Richard Socher. 2019. Neural text summarization: A critical evaluation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 540–551.
- Lewis et al. (2020) Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov, and Luke Zettlemoyer. 2020. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.
- Li et al. (2018) Haoran Li, Junnan Zhu, Jiajun Zhang, and Chengqing Zong. 2018. Ensure the correctness of the summary: Incorporate entailment knowledge into abstractive sentence summarization. In Proceedings of the 27th International Conference on Computational Linguistics, pages 1430–1441.
- Manning et al. (2014) Christopher D. Manning, Mihai Surdeanu, John Bauer, Jenny Finkel, Steven J. Bethard, and David McClosky. 2014. The Stanford CoreNLP natural language processing toolkit. In Association for Computational Linguistics (ACL) System Demonstrations, pages 55–60.
- Mao et al. (2019) Huanru Henry Mao, Bodhisattwa Prasad Majumder, Julian McAuley, and Garrison Cottrell. 2019. Improving Neural Story Generation by Targeted Common Sense Grounding. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP).
- Maynez et al. (2020) Joshua Maynez, Shashi Narayan, Bernd Bohnet, and Ryan McDonald. 2020. On Faithfulness and Factuality in Abstractive Summarization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.
- Mel’čuk (1988) Igor Mel’čuk. 1988. Dependency Syntax: Theory and Practice. State University of New York Press.
- Michael et al. (2018) Julian Michael, Gabriel Stanovsky, Luheng He, Ido Dagan, and Luke Zettlemoyer. 2018. Crowdsourcing Question-Answer Meaning Representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers), pages 560–568.
- Pasunuru and Bansal (2018) Ramakanth Pasunuru and Mohit Bansal. 2018. Multi-Reward Reinforced Summarization with Saliency and Entailment. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers).
- Pavlick and Kwiatkowski (2019) Ellie Pavlick and Tom Kwiatkowski. 2019. Inherent Disagreements in Human Textual Inferences. Transactions of the Association for Computational Linguistics.
- Pennington et al. (2014) Jeffrey Pennington, Richard Socher, and Christopher D. Manning. 2014. GloVe: Global Vectors for Word Representation. In Empirical Methods in Natural Language Processing (EMNLP).
- Radford et al. (2019) Alec Radford, Jeff Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language Models are Unsupervised Multitask Learners.
- Shen et al. (2020) Xiaoyu Shen, Ernie Chang, Hui Su, Jie Zhou, and Dietrich Klakow. 2020. Neural Data-to-Text Generation via Jointly Learning the Segmentation and Correspondence. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.
- Song et al. (2018) Kaiqiang Song, Lin Zhao, and Fei Liu. 2018. Structure-Infused Copy Mechanisms for Abstractive Summarization. In Proceedings of the 27th International Conference on Computational Linguistics, pages 1717–1729.
- Wang et al. (2020) Alex Wang, Kyunghyun Cho, and Mike Lewis. 2020. Asking and Answering Questions to Evaluate the Factual Consistency of Summaries. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.
- White et al. (2016) Aaron Steven White, Drew Reisinger, Keisuke Sakaguchi, Tim Vieira, Sheng Zhang, Rachel Rudinger, Kyle Rawlins, and Benjamin Van Durme. 2016. Universal decompositional semantics on universal dependencies. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 1713–1723.
- Wieting and Gimpel (2018) John Wieting and Kevin Gimpel. 2018. ParaNMT-50M: Pushing the limits of paraphrastic sentence embeddings with millions of machine translations. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 451–462, Melbourne, Australia. Association for Computational Linguistics.
- Williams et al. (2018) Adina Williams, Nikita Nangia, and Samuel Bowman. 2018. A broad-coverage challenge corpus for sentence understanding through inference. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers). Association for Computational Linguistics.
- Wolf et al. (2019) Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, R’emi Louf, Morgan Funtowicz, and Jamie Brew. 2019. Huggingface’s transformers: State-of-the-art natural language processing. ArXiv, abs/1910.03771.
- Yu et al. (2018) Adams Wei Yu, David Dohan, Quoc Le, Thang Luong, Rui Zhao, and Kai Chen. 2018. QANet: Combining Local Convolution with Global Self-Attention for Reading Comprehension. In International Conference on Learning Representations.
- Zhang and Bansal (2019) Shiyue Zhang and Mohit Bansal. 2019. Addressing Semantic Drift in Question Generation for Semi-Supervised Question Answering. In Empirical Methods in Natural Language Processing (EMNLP).
- Zhang et al. (2020) Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian Q. Weinberger, and Yoav Artzi. 2020. BERTScore: Evaluating Text Generation with BERT. In International Conference on Learning Representations.
- Zhang et al. (2019) Yuan Zhang, Jason Baldridge, and Luheng He. 2019. PAWS: Paraphrase Adversaries from Word Scrambling. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers).
- Zhou and Bansal (2020) Xiang Zhou and Mohit Bansal. 2020. Towards Robustifying NLI Models Against Lexical Dataset Biases. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.
Appendix A Dependency Label Set
As outlined in Section 2, we restrict our analysis to a subset of dependency arcs which are more strongly connected to semantics. For each of hypothesis corresponding to the input , we extract Enhanced Dependencies using the Stanford CoreNLP tool, and assign entailment labels to this dependency arc set using the strategy outlined in Section 4. We exclude the following arcs from our analysis: punct, det, case, aux, auxpass, dep, cop, mark. This same subset of arcs are ignored while computing sentence-level factuality.
Appendix B Examples from Synonym Test Set
As outlined in Section 5.2, we additionally augment our auto-derived training data (AD) with synonym data (S) and show that this improves the model performance on the held out synonym only test set. Figure 6 provides some examples showing the predicted entailment probability for each arc using this augmented training data. The predictions show that our model learns some bias to recognize synonym replacements and small phrasal substitutions as arcs that are entailed by the input.

Appendix C Implementation Details
To train our DAE model, we fine-tune the pre-trained encoders Bert (bert-base-uncased, 110M parameters) and Electra (electra-base-discriminator, 110M parameters), as outlined in Section 5. We perform hyperparameter trials, varying only the learning rate using manual tuning. The models with the best dev set accuracy are used. The final hyperparameters used are:
Implementation Library | transformers Wolf et al. (2019) |
---|---|
Computing Infrastructure | 32GB NVIDIA V100 GPU |
Max Seq Length | 128 |
Linear Layer Size | (2304, 2) |
Optimizer | Adam |
Optimizer Params | |
Learning Rate Decay | Linear |
Learning rate | 1e-5 |
Weight Decay | 0 |
Warmup Steps | 0 |
Maximum Gradient Norm | 1 |
Batch size | 32 |
Epochs | 3 |
Additionally, we fine-tune (bert-base-uncased, 110M parameters) and Electra (electra-base-discriminator, 110M parameters) models on the MNLI dataset. We fine-tuned the model using hyperparameter trials, varying only the learning rate using manual tuning. The models with the best dev set accuracy are used. The final hyperparameters used are shown in Table 8.
Implementation Library | transformers Wolf et al. (2019) |
---|---|
Computing Infrastructure | 32GB NVIDIA V100 GPU |
Max Seq Length | 256 |
Linear Layer Size | (768, 2) |
Optimizer | Adam |
Optimizer Params | |
Learning Rate Decay | Linear |
Learning rate | 2e-5 |
Weight Decay | 0 |
Warmup Steps | 0 |
Maximum Gradient Norm | 1 |
Batch size | 32 |
Epochs | 3 |
We get a dev accuracy of and for the Bert and Electra models respectively.