Entity-level Factual Consistency of Abstractive Text Summarization
Abstract
A key challenge for abstractive summarization is ensuring factual consistency of the generated summary with respect to the original document. For example, state-of-the-art models trained on existing datasets exhibit entity hallucination, generating names of entities that are not present in the source document. We propose a set of new metrics to quantify the entity-level factual consistency of generated summaries and we show that the entity hallucination problem can be alleviated by simply filtering the training data. In addition, we propose a summary-worthy entity classification task to the training process as well as a joint entity and summary generation approach, which yield further improvements in entity level metrics.
1 Introduction
Many recent advances in deep neural networks have led to significant improvement in the quality of abstractive summarization Radford et al. (2019); Gehrmann et al. (2019); Lewis et al. (2019). Despite this progress, there are still many limitations facing neural text summarization Kryscinski et al. (2019), the most serious of which is their tendency to generate summaries that are not factually consistent with the input document; a factually consistent summary only contains statements that can be derived from the source document. Recent studies show that about 30% of the summaries generated by neural network sequence-to-sequence models suffer from fact fabrication Cao et al. (2018). Unfortunately, the widely used ROUGE score is inadequate to quantify factual consistency Kryscinski et al. (2019).
Factual inconsistency can occur at either the entity or the relation level. At the entity level, a model generated summary may contain named-entities that never appeared in the source document. We call this the entity hallucination problem. For example, consider the following model generated summary:
People in Italy and the Netherlands are more likely to consume fewer cups of coffee than those in the UK, a study suggests.
“UK” never appeared in the input source document (taken from the test set of the XSUM dataset Narayan et al. (2018)). In fact, the source document mentioned a study involving people in Italy and Netherlands; “UK” was a result of model hallucination. Another type of inconsistency occurs when the entities indeed exist in the source document but the relations between them are not in the source document. This type of inconsistency is much harder to identify. Open Information Extraction (OpenIE) and dependency parsing tools have been used Cao et al. (2018) to identify the underlying relations in a summary, but are not yet accurate enough for practical use. Ultimately, these researchers relied on manually classifying generated summaries into faithful, fake, or unclear.
In this paper, we propose a set of simple metrics to quantify factual consistency at the entity-level. We analyze the factual quality of summaries produced by the state-of-the-art BART model Lewis et al. (2019) on three news datasets. We then propose several techniques including data filtering, multi-task learning and joint sequence generation to improve performance on these metrics. We leave the relation level consistency to future work.
2 Related work
Large transformer-based neural architectures combined with pre-training have set new records across many natural language processing tasks Vaswani et al. (2017); Devlin et al. (2019); Radford et al. (2019). In particular, the BART model Lewis et al. (2019) has shown superior performance in many text generation tasks including abstractive summarization. In contrast to encoder-only pre-training such as in BERT Devlin et al. (2019) or decoder-only pre-training such as in GPT-2 Radford et al. (2019), BART is an encoder-decoder transformer-based neural translation model jointly pre-trained to reconstruct corrupted input sequences of text.
Several authors have pointed out the problem of factual inconsistency in abstractive summarization models Kryscinski et al. (2019); Kryściński et al. (2019); Cao et al. (2018); Welleck et al. (2019). The authors in Kryściński et al. (2019) proposed to train a neural network model to classify if a summary is factually consistent with a given source document, similar to a natural language inference task. In the dialogue generation setting, authors in Li et al. (2019) proposed using unlikelihood to surpress logically inconsistent responses. Our work is complementary to such existing approaches as we focus on simple entity-level metrics to quantify and improve factual consistency. Our goal of improving entity-level metrics of summaries is also related to controllable abstractive summarization Fan et al. (2018), where a list of named-entities that a user wants to see in the summary can be passed as input to influence the generated summary. In contrast, our goal is to predict which entities are summary-worthy while generating the summary that contains them. In this view we are trying to solve a more challenging problem.
3 Entity-level factual consistency metrics
We propose three new metrics that rely on off-the-shelf tools to perform Named-Entity Recognition (NER). 111We use Spacy Honnibal and Montani (2017). We use and to denote the number of named-entities in the target (gold summary) and hypothesis (generated summary), respectively. We use to denote the number of entities found in the generated summary that can find a match in the source document. If a named-entity in the summary consists of multiple words, we consider it a match as long as any n-gram of the named-entity can be found in the source document. This is meant to capture the situation where the named-entity can be shortened; for example, “Obama ” is a match for “Barack Obama” and “Harvard” is a match for “Harvard University”. When the match is at the unigram level, we make sure that it is not a stop word such as “the”. We also make the match case-insensitive to accommodate casing variances.
Precision-source:
We propose precision-source () to quantify the degree of hallucination with respect to the source: It is simply the percentage of named-entities in the summary that can be found in the source. Low means hallucination is severe.
We first evaluate the score on the ground truth summaries of the 3 datasets: Newsroom Grusky et al. (2018), CNN/DailyMail Nallapati et al. (2016) and XSUM Narayan et al. (2018).
Newsroom | CNNDM | XSUM | |||||||
train | val | test | train | val | test | train | val | test | |
avg. | 2.08 | 2.10 | 2.09 | 4.36 | 5.09 | 4.87 | 2.08 | 2.06 | 2.08 |
avg. | 1.88 | 1.90 | 1.90 | 4.21 | 4.92 | 4.70 | 1.64 | 1.64 | 1.64 |
(%) | 90.6 | 90.6 | 90.5 | 96.5 | 96.7 | 96.6 | 79.0 | 79.5 | 79.3 |
Table 1 shows that among the three datasets, the ground truth summaries in XSUM have the lowest score. This is because the ground truth summaries in the XSUM dataset often use the first sentence of the article as the summary; the source document is constructed to be the rest of the article and may not repeat the named-entities that appeared in the summary. We hypothesize that the hallucination problem is largely caused by the training data itself. Thus, we propose to perform entity-based data filtering to construct a “clean” version of these datasets as described next.
Entity-based data filtering:
For each dataset, we apply Spacy NER on the gold summary to identify all the named-entities. 222We ignore certain types of entities such as date, time, numerals because they tend to have large variations in representation and are difficult to determine a match in the source document. The appendix contains more details. If any of the entities cannot find a match in the source document, we discard the sentence that contains the entity from the ground truth summary. If the ground truth summary consists of only one sentence and it needs to be discarded, we remove the document-summary pair from the dataset. This way, we ensure that our filtered dataset does not contain hallucination of entities () in the ground truth summary. The dataset size before and after the filtering is shown in Table 2. About a third of examples are filtered out for XSUM. Again, this is because of the way XSUM dataset is constructed as mentioned in the previous paragraph. As we shall see in Table 3, entity-based data filtering reduces hallucination of the trained model and the effect is especially significant in the XSUM dataset.
Newsroom | CNNDM | XSUM | |||||||
train | val | test | train | val | test | train | val | test | |
original | 922,500 (1.58) | 100,968 (1.60) | 100,933 (1.59) | 287,112 (3.90) | 13,368 (4.13) | 11,490 (3.92) | 203,540 (1.0) | 11,301 (1.0) | 11,299 (1.0) |
---|---|---|---|---|---|---|---|---|---|
after filtering | 855,975 (1.62) | 93,678 (1.64) | 93,486 (1.64) | 286,791 (3.77) | 13,350 (3.99) | 11,483 (3.77) | 135,155 (1.0) | 7,639 (1.0) | 7,574 (1.0) |
Precision-target and recall-target:
Although the precision-source () metric quantifies the degree of entity hallucination with respect to the source document, it does not capture the entity-level accuracy of the generated summary with respect to the ground truth summary. To get a complete picture of the entity-level accuracy of the generated summary, we propose the precision-target () score: where is the number of named-entities in the generated summary that can find a match in the ground truth summary; and the recall-target () score: where is the number of named-entities in the ground truth summary. We compute the F1 score as .
training data | Rouge1 | Rouge2 | RougeL | macro | micro | macro | micro | macro | micro | macro | micro | |
---|---|---|---|---|---|---|---|---|---|---|---|---|
Newsroom | original | 47.70.2 | 35.00.3 | 44.10.2 | 97.20.1 | 97.00.1 | 65.40.3 | 62.90.4 | 70.80.3 | 68.5 0.2 | 68.00.2 | 65.60.3 |
+ filtering | 47.70.1 | 35.10.1 | 44.1 0.1 | 98.10.1 | 98.00.0 | 66.50.1 | 63.80.1 | 70.2 0.2 | 67.70.3 | 68.30.1 | 65.70.1 | |
+ classification | 47.70.2 | 35.10.1 | 44.20.2 | 98.10.1 | 98.00.0 | 67.20.4 | 64.20.4 | 70.30.2 | 67.80.4 | 68.70.3 | 65.90.4 | |
JAENS | 46.6 0.5 | 34.30.3 | 43.20.3 | 98.30.1 | 98.30.1 | 69.51.6 | 67.31.2 | 68.91.5 | 66.81.6 | 69.20.1 | 67.00.2 | |
CNNDM | original | 43.70.1 | 21.10.1 | 40.60.1 | 99.50.1 | 99.40.1 | 66.00.4 | 66.50.4 | 74.70.7 | 75.40.6 | 70.00.2 | 70.70.3 |
+ filtering | 43.40.2 | 20.80.1 | 40.30.2 | 99.90.0 | 99.90.0 | 66.2 0.4 | 66.60.3 | 74.10.6 | 74.90.6 | 69.90.2 | 70.50.2 | |
+ classification | 43.50.2 | 20.80.2 | 40.40.2 | 99.90.0 | 99.90.0 | 67.00.6 | 67.50.5 | 74.70.2 | 75.50.1 | 70.60.3 | 71.30.3 | |
JAENS | 42.4 0.6 | 20.20.2 | 39.50.5 | 99.90.0 | 99.90.0 | 67.90.7 | 68.40.6 | 75.10.7 | 76.40.7 | 71.30.2 | 72.20.2 | |
XSUM | original | 45.60.1 | 22.50.1 | 37.20.1 | 93.90.1 | 93.60.2 | 74.10.2 | 73.30.2 | 80.10.1 | 80.30.3 | 77.00.1 | 76.60.2 |
+ filtering | 45.40.1 | 22.20.1 | 36.90.1 | 98.20.0 | 98.20.1 | 77.90.2 | 77.30.2 | 79.40.2 | 79.60.2 | 78.60.1 | 78.40.2 | |
+ classification | 45.30.1 | 22.10.0 | 36.90.1 | 98.30.1 | 98.20.1 | 78.60.3 | 78.00.3 | 79.50.3 | 79.80.4 | 79.10.1 | 78.90.1 | |
JAENS | 43.40.7 | 21.00.3 | 35.5 0.4 | 99.00.1 | 99.00.1 | 77.60.9 | 77.10.6 | 79.50.6 | 80.00.5 | 78.50.2 | 78.50.1 |
4 Multi-task learning:
In addition to entity-based data filtering, we also explore another method to further improve the summarization quality. In particular, we incorporate an additional task of classifying summary-worthy named-entities in the source document. A summary-worthy named-entity in the source document is one that appears in the ground truth summary and thus, is a salient entity, worthy of inclusion in the generated summary. Intuitively, if we can identify these summary-worthy named-entities using the encoder representation, we may potentially increase the entity-level precision and recall metrics as well as the overall quality of the summary. We achieve this by adding a classification head to the encoder of BART. To prepare for the classification label, we first identify the named-entities in the ground truth summary and find the matching tokens in the source document. We then assign the (B)eginning-(I)nside-(O)utside labels to each token of the source document to denote if the token is beginning, inside or outside of a summary-worthy named-entity, respectively. During training, we simply add the classification loss for each token at the encoder to the original sequence-to-sequence loss.
More precisely, let be a dataset of examples where are the tokens of the th source document and are the tokens of the target (ground truth summary). The standard sequence-to-sequence training minimizes the maximum log likelihood estimation (MLE) loss:
With summary-worthy entity classification, each example has an additional sequence of BIO labels . By adding an additional fully connected layer on top of the BART encoder, we obtain the classification loss
Finally, we can minimize the joint loss where is a hyper parameter. We choose between 0.1 to 0.5 via the validation sets.
5 Joint Entity and Summary Generation:
We also explore another generative approach to promote entity-level precision and recall metrics. In particular, instead of just generating the summary, we train the BART model to generate the sequence of summary-worthy named-entities, followed by a special token, and then the summary. We call this approach JAENS (Join sAlient ENtity and Summary generation). Similar to the multi-task learning approach discussed earlier, JAENS encourages the model to jointly learn to identify the summary-worthy named-entities while learning to generate summaries. Since the decoder generates the salient named-entities first, the summaries that JAENS generate can further attend to these salient named-entities through decoder self-attention.
6 Experiment results
We use the pre-trained BART-large model in the Fairseq library Ott et al. (2019) to fine-tune on the 3 summarization datasets.333Our code is available at https://github.com/amazon-research/fact-check-summarization The appendix contains additional details of experimental setup.
In Table 3, we show the effect of the entity-based data filtering. For each dataset, we train two separate models: using the training data before and after entity-based data filtering as shown in Table 2. We evaluate both models on the “clean” test set after entity-based data filtering. We choose this filtered version of the original test set because we only want to measure entity-level consistency against the correct set of entities; using the unfiltered dataset means we could count a hallucinated entity as correct. We observe improvements of across all three datasets trained using the filtered subset of data. For example in XSUM, the is increased from 93.6% to 98.2%, indicating a significant reduction in entity hallucination. In addition, the entity-based data filtering generally improves other entity-level metrics as well. Even with less training data, the entity-based data filtering is able to maintain the ROUGE scores quite well. For XSUM, about 34% of the training data is filtered out (c.f. Table 2), which explains the more noticable impact on the ROUGE scores. The results in Table 3 suggest that entity-level data filtering is a simple yet effective approach to achieve higher entity-level factual consistency as well as general summarization quality. In Table 4 we provide qualitative examples where the model trained on the original data produces hallucination and the entity-level data filtering removes such hallucination.
Table 3 shows that adding the classification task (multi-task) futher increases the and metric and therefore the overall entity-level on top of the improvements from data filtering. Similar gains can be observed with JAENS, which out-performs the multi-task approach on CNNDM and Newsroom datasets. The result confirms our intuition that the summaries in JAENS can benefit from attending to the generated salient entities in terms of the entity level metrics. However, the additional complexity during decoding may have hurt the ROUGE scores.
For the interested readers, we also evaluated the PEGASUS Zhang et al. (2020) models for the ROUGE and entity level metrics on these three datasets in the appendix.
Accuracy of entity level metrics:
As our entity level metrics are based on automatic NER tools and heuristics matching rules, errors in both steps can lead to inaccuracy in the metrics. By manually checking 10 random ground truth summaries together with the source documents in the validation split of XSUM dataset, we found that all of the named entities are correctly identified by the NER tool and the matchings are correct. Therefore, we believe that even our current NER tool and matching rule already produce high accuracy in practice.
Before data filtering | After data filtering | With classification | Ground truth summary |
---|---|---|---|
People in Italy and the Netherlands are more likely to consume fewer cups of coffee than those in the UK, a study suggests. | The desire to drink coffee may be encoded in our DNA, according to scientists. | People with a particular gene are more likely to consume fewer cups of coffee, a study has suggested. | Researchers have identified a gene that appears to curb coffee consumption. |
A cathedral in Surrey is set to be restored after more than £5m was raised to pay for repairs and improvements. | A £7m project to save a Grade II-listed cathedral from demolition is set to go ahead. | A cathedral which has been threatened with demolition is set to be saved by a £5m fundraising campaign. | A 1960s-built cathedral that was ”at serious risk of closure” has raised more than 90% of its £7m target for urgent repairs and development. |
More than 800,000 chemists in the Indian capital, Delhi, have gone on strike in protest against online drug sales. | More than 800,000 chemists in India will go on strike on Wednesday to protest against illegal online drug sales. | More than 800,000 chemists in India are set to go on strike on Wednesday in a row over the sale of drugs online. | At least 800,000 pharmacies in India are on a one-day strike, demanding an end to online drug sales which they say is affecting their business. |
Police officers in Pembrokeshire are to be issued with body-worn cameras. | Police officers in Powys are to be issued with body-worn cameras in a bid to improve transparency in the force. | Police officers in Powys are to be issued with body cameras in a bid to improve transparency in the force. | A police force has begun the rollout of body cameras for 800 officers and community support officers. |
Wales midfielder Becky Lawrence has been speaking to BBC Sport about her time as a player-manager with Melbourne City. | It’s been a great few weeks for me as a player-manager and now I’m heading home to Wales ahead of the Cyprus Cup. | It’s been a very busy few weeks for me as I’m heading home to Wales ahead of the Cyprus Cup. | I have certainly had worse 24 hours in my life than winning the Grand Final with Melbourne City and then being named in the Wales squad for the Cyprus Cup. |
7 Conclusion
In this paper we study the entity-level factual consistency of the state-of-the-art summarization model. We propose precision-source score to quantify the degree of entity hallucination. We also propose additional metrics and to measure entity level accuracy of the generated summary with respect to the ground truth summary. We found that the ground truth summaries of the XSUM dataset contain a high level of entity hallucination. We propose a simple entity-level data filtering technique to remove such hallucination in the training data. Experiments show that such data filtering leads to significant improvement in . ( increases from below 94% to above 98% in XSUM for example.) We futher proposed a multi-task learning and a joint sequence generation approach to further improve the entity-level metrics. Overall, combining our proposed approaches significantly reduces entity hallucination and leads to higher entity level metrics with minimal degradation of the ROUGE scores.
References
- Cao et al. (2018) Ziqiang Cao, Furu Wei, Wenjie Li, and Sujian Li. 2018. Faithful to the original: Fact aware neural abstractive summarization.
- Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
- Fan et al. (2018) Angela Fan, David Grangier, and Michael Auli. 2018. Controllable abstractive summarization. In Proceedings of the 2nd Workshop on Neural Machine Translation and Generation, pages 45–54, Melbourne, Australia. Association for Computational Linguistics.
- Gehrmann et al. (2019) Sebastian Gehrmann, Zachary Ziegler, and Alexander Rush. 2019. Generating abstractive summaries with finetuned language models. In Proceedings of the 12th International Conference on Natural Language Generation, pages 516–522, Tokyo, Japan. Association for Computational Linguistics.
- Grusky et al. (2018) Max Grusky, Mor Naaman, and Yoav Artzi. 2018. Newsroom: A dataset of 1.3 million summaries with diverse extractive strategies. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 708–719, New Orleans, Louisiana. Association for Computational Linguistics.
- Honnibal and Montani (2017) Matthew Honnibal and Ines Montani. 2017. spaCy 2: Natural language understanding with Bloom embeddings, convolutional neural networks and incremental parsing. To appear.
- Kryściński et al. (2019) Wojciech Kryściński, Bryan McCann, Caiming Xiong, and Richard Socher. 2019. Evaluating the factual consistency of abstractive text summarization.
- Kryscinski et al. (2019) Wojciech Kryscinski, Nitish Shirish Keskar, Bryan McCann, Caiming Xiong, and Richard Socher. 2019. Neural text summarization: A critical evaluation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 540–551, Hong Kong, China. Association for Computational Linguistics.
- Lewis et al. (2019) Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Veselin Stoyanov, and Luke Zettlemoyer. 2019. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461.
- Li et al. (2019) Margaret Li, Stephen Roller, Ilia Kulikov, Sean Welleck, Y-Lan Boureau, Kyunghyun Cho, and Jason Weston. 2019. Don’t say that! making inconsistent dialogue unlikely with unlikelihood training. CoRR, abs/1911.03860.
- Nallapati et al. (2016) Ramesh Nallapati, Bowen Zhou, Cicero dos Santos, Çağlar Gu̇lçehre, and Bing Xiang. 2016. Abstractive text summarization using sequence-to-sequence RNNs and beyond. In Proceedings of The 20th SIGNLL Conference on Computational Natural Language Learning, pages 280–290, Berlin, Germany. Association for Computational Linguistics.
- Narayan et al. (2018) Shashi Narayan, Shay B. Cohen, and Mirella Lapata. 2018. Don’t give me the details, just the summary! topic-aware convolutional neural networks for extreme summarization. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 1797–1807, Brussels, Belgium. Association for Computational Linguistics.
- Ott et al. (2019) Myle Ott, Sergey Edunov, Alexei Baevski, Angela Fan, Sam Gross, Nathan Ng, David Grangier, and Michael Auli. 2019. fairseq: A fast, extensible toolkit for sequence modeling. In Proceedings of NAACL-HLT 2019: Demonstrations.
- Radford et al. (2019) Alec Radford, Jeff Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language models are unsupervised multitask learners.
- Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems 30, pages 5998–6008. Curran Associates, Inc.
- Welleck et al. (2019) Sean Welleck, Jason Weston, Arthur Szlam, and Kyunghyun Cho. 2019. Dialogue natural language inference. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 3731–3741, Florence, Italy. Association for Computational Linguistics.
- Wolf et al. (2020) Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander M. Rush. 2020. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online. Association for Computational Linguistics.
- Zhang et al. (2020) Jingqing Zhang, Yao Zhao, Mohammad Saleh, and Peter Liu. 2020. PEGASUS: Pre-training with extracted gap-sentences for abstractive summarization. In Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 11328–11339. PMLR.
Rouge1 | Rouge2 | RougeL | macro | micro | macro | micro | macro | micro | macro | micro | |
---|---|---|---|---|---|---|---|---|---|---|---|
Newsroom | 40.6 | 28.4 | 37.4 | 94.6 | 94.7 | 53.4 | 55.5 | 68.5 | 67.8 | 60.0 | 61.1 |
CNNDM | 42.5 | 20.7 | 39.6 | 99.1 | 99.0 | 65.9 | 66.7 | 74.7 | 75.7 | 70.0 | 70.9 |
XSUM | 45.3 | 23.7 | 37.9 | 93.9 | 93.1 | 76.6 | 75.8 | 80.3 | 80.1 | 78.4 | 77.9 |
Appendix A Supplementary material for Entity-level Factual Consistency of Abstractive Text Summarization
A.1 Details of NER filtering
We only consider named-entities of the following types: ’PERSON’ (People, including fictional.), ’FAC’ (Buildings, airports, highways, bridges, etc.), ’GPE’ (Countries, cities, states.), ’ORG’ (Companies, agencies, institutions, etc.), ’NORP’ (Nationalities or religious or political groups.), ’LOC’ (Non-GPE locations, mountain ranges, bodies of water.), ’EVENT’ (Named hurricanes, battles, wars, sports events, etc.). We ignore other types of entities such as date, time, numerals because they tend to have large variations in representation and are difficult to determine a match in the source document.
A.2 Details of experimental setup
We use the pre-trained BART-large model in the Fairseq library Ott et al. (2019) to fine-tune on the 3 summarization datasets.
In all experiments, we validate the ROUGE scores of the generated summaries on the validation split and early-stop on the epoch with the highest validation score. We use the standard learning rate of 3e-5 for finetuning with linear decay schedule and 500 warmup steps. For Newsroom, we use 4 p3.16xlarge EC2 instances on AWS with a total of 32 Tesla V100 GPUs for finetuning and the effective batch size is 32; for XSUM, we use 1 p3.16xlarge instance with a total of 8 Tesla V100 GPUs and update frequency of 4, giving an effective batch size of 32; for CNNDM, we use 1 p3.16xlarge instance with a total of 8 Tesla V100 GPUs, giving an effective batch size of 8.
We chose the parameter for multi-task learning between 0.1 and 0.5 with step of 0.05 based on ROUGE scores on the validation set. We found the best values are 0.3, 0.3 and 0.15 for Newsroom, CNNDM and XSUM, respectively. We observe that the ROUGE and entity level metrics on validation and test sets are very close, with the former slightly higher.
During decoding, we use beam size of 1 for Newsroom, 4 for CNNDM and 6 for XSUM (to be consistent with the setting in Lewis et al. (2019)). We did use trigrams blocking in beam search as we did not see much need for this additional step.
A.3 Evaluation of PEGASUS Zhang et al. (2020)
In this section we simply evaluate the PEGASUS checkpoints provided by Huggingface Wolf et al. (2020) on the NER filtered test sets. The checkpoints are downloaded from https://huggingface.co/google/pegasus-newsroom, https://huggingface.co/google/pegasus-cnn_dailymail and https://huggingface.co/google/pegasus-xsum, respectively. The results are summarized in Table 5. Note that PEGASUS performances similarly on CNNDM and XSUM but worse on Newsroom compared to BART-large.