Fill in the BLANC:
Human-free quality estimation
of document summaries
Abstract
We present BLANC, a new approach to the automatic estimation of document summary quality. Our goal is to measure the functional performance of a summary with an objective, reproducible, and fully automated method. Our approach achieves this by measuring the performance boost gained by a pre-trained language model with access to a document summary while carrying out its language understanding task on the document’s text. We present evidence that BLANC scores have as good correlation with human evaluations as do the ROUGE family of summary quality measurements. And unlike ROUGE, the BLANC method does not require human-written reference summaries, allowing for fully human-free summary quality estimation.
1 Introduction
Two most widely used methods for measuring the quality of a summary are ROUGE Lin (2004) and human evaluation Kryściński et al. (2019a).
The ROUGE family of methods are well-defined and reproducible. However, these methods typically require a human-written reference summaries for comparison, completely disregarding the original document text. Even if one assumes that a reference summary is available and of optimal quality, the ROUGE method is limited to measuring a mechanical overlap of text tokens with little regard to semantics. This deficiency may be partially addressable through measurement of the similarity not of text tokens but named entities or other preprocessed features Mao et al. (2019); Cohan and Goharian (2016); Elghannam and El-Shishtawy (2015); Ng and Abrecht (2015); Ganesan (2018) or embeddings Zhao et al. (2019); Zhang et al. (2020); Gao et al. (2020). In the latter work Gao et al. (2020) the references are not human-written but unsupervisedly constructed from selected salient sentences. An overlap can be measured as well between summary and document text Shao et al. (2017).
Human evaluation of summary quality is far more meaningful and powerful than ROUGE, but it is far less reproducible. Summary quality estimation is a cognitively demanding and highly subjective task. Humans are also vulnerable to biases, such as the preference for phrases and sentences copied directly from the document text into summaries Ziegler et al. (2020). Improving human evaluation may require prompting labelers to pay higher attention Hardy et al. (2019), as well as splitting quality scores into multiple dimensions such as fluency, informativeness, and factual correctness Kryściński et al. (2019a, b); Fan et al. (2018). Even if humans can be trained to be more reliable, reproducible estimators of summary quality, they will forever remain a slow, expensive, limiting resource.
One possible route to a better automatic method for summary quality estimation is to train a model on document summaries annotated with human quality scores Louis and Nenkova (2009, 2013); Xenouleas et al. (2019). Such a model could be used to evaluate summaries without further human involvement. But even if such a model could achieve high agreement with human labelers, its performance would only be as objective and reproducible as the summary quality scores generated by one particular group of humans on a particular group of documents. Such a model may not generalize beyond the domain and style of the training samples unless they are a massive, representative sample of all documents of interest.
A more fundamental approach to the problem is to estimate how ”helpful” a summary is for the task of understanding a text. For example this might be achieved through a series of question-answers Eyal et al. (2019); Chen et al. (2018); Scialom et al. (2019). However, with this approach one must choose from a vast set of questions one might ask of a text, presupposing knowledge of the document itself and seriously limiting its reproducibility.
In the following section we suggest a new approach that is fundamentally justifiable as an estimator of summary quality, as well as being conceptually simple and reproducible.
2 Methods
2.1 Introducing BLANC
An ideal estimator should directly test how helpful a summary is to its readers. It should reliably estimate quality across a broad range of document domains and styles. And yet it should achieve this without requiring ornate preconditions and presuppositions about the text being summarized. If this estimator relies upon an existing base model, that model should be well-documented, well-understood and widely used.
We propose BLANC111According to ancient tradition, we should adorn our newly created jargon term with a bacronymic justification. The term BLANC is a nod to its proud lineage of French color words that began with the BLEU method for evaluating machine translation and ROUGE for summarization. BLANC is also a reference to the method’s core task of ”filling in the blanks” in the masked token task. But to honor tradition we offer this: Bacronymic Language model Approach for summary quality estimatioN. Cool? as a replacement for the ROUGE family of summary quality estimators.
We define BLANC as a measure of how well a summary helps an independent, pre-trained language model while it performs its language understanding task on a document. We focus on the masked token task, also known as the Cloze task Taylor (1953), in which a model is challenged to reconstruct obscured spans of text. We use the well-known BERT language model Devlin et al. (2018) pre-trained to predict masked text tokens (words or sub-words). The BERT tokenizer represents the majority of the most frequently used words as single tokens, while splitting less common words into two or more.
We present two versions of BLANC, which we dub BLANC-help and a BLANC-tune. These measures are described in detail in the following sections. The essential difference between them:
-
1.
BLANC-help uses the summary text by directly concatenating it to each document sentence during inference.
-
2.
BLANC-tune uses the summary text to fine-tune the language model, and then processes the entire document.
Thus with BLANC-help, the language model refers to the summary each time it attempts to understand a part of the document text. While with BLANC-tune, the model learns from the summary first, and then uses its gained skill to help it understand the entire document.
2.2 BLANC-help
The algorithm for obtaining BLANC-help scores is illustrated in Figure 1.

There are many possible choices for how to mask the tokens. Our aim is to evenly cover all tokens in a sentence with a certain frequency, for the sake of full reproducibility. A random coverage is also possible, but it is not as conveniently reproducible.
The unmasking is done twice for each sentence of the text and for each allowed choice of masked tokens in the sentence. First, the unmasking is done for input composed of the summary concatenated with the sentence. Second, the unmasking is done for input composed of a ”filler” concatenated with the sentence. The filler has exactly the same lengths as the summary, but each summary token is replaced by a period symbol (”.”). After iterating over all sentences and over all the allowed choices of masking, we end up with four total counts of successful and unsuccessful unmasking . Here the index equals 0 or 1 - for unsuccessful (0) or successful (1) unmasking for the filler-input. The index is defined the same way for the summary-input. For example, is the count of cases where the filler-input was unsuccessful and the summary-input was successful.
We define BLANC-help as the difference between the accuracy of unmasking with the summary and the accuracy of unmasking with the filler:
The accuracies are and . The total count is . The BLANC value can range from -1 to 1, but as shown in next sections the typical values are between 0 (summary is useless) and 0.3 (summary provides 30% help).
The algorithm for BLANC-help is shown in more detail in Figure 2.
Given: ; ; ; |
Parameters , |
Initialise |
Initialise , , , to zero |
for in : |
for in range from to : |
Mask each th word if |
and if |
for each position in masked tokens: |
Since the BERT model deals with tokens rather than words, we can choose to mask tokens rather than words. In typical news documents only about 10% of words are split by the BERT tokenizer into two or more tokens. Such ”composite” words (not existing in the BERT vocabulary) should be particularly valuable in estimating the helpfulness of a summary. In a version dealing with tokens rather than words it is natural to always allow masking of composite words regardless of their length.
The setting allows the masking only of sufficiently long words (4 or more characters), because shorter words are typically easier to predict, with or without the help of a summary.
The value in Figure 2 is a natural choice because the standard BERT model is trained by masking 15% of tokens, which makes about one-sixth of tokens eligible to be masked.
We found that altering the filler has a negligible effect on the measures. The reason we use the filler is to avoid any effect of the length of input on the action of the model.
2.3 BLANC-tune
The algorithm for obtaining BLANC-tune is illustrated in Figure 3.

For calculating this measure, the model first learns from the summary, and then we observe how helpful this learning was in reconstructing masked tokens in text sentences.
As in the case of BLANC-help, we define BLANC-tune by comparing the accuracy of two reconstructions: one that does use the summary, and another that does not. In the case of BLANC-help, this was the difference between placing the summary vs. placing the filler in front of a sentence. Now, in the case of BLANC-tune, we compare the performance of a model fine-tuned on the summary text vs. a model that has never seen the summary.
The task, using a model to unmask tokens, is performed the same way as for BLANC-help, except that the input is simply a document sentence with masked tokens.
The tuning of the model is done on an extremely small dataset (derived from the summary text), in which each sample is the very same summary but with different tokens masked. The masking in the summary is done according to the original BERT pre-training strategy. Unmasking must be performed for 15% randomly selected tokens, of which 80% are masked, 10% are replaced by random tokens, and 10% are left unchanged. To ensure coverage of tokens, we select and shuffle all eligible tokens, and then go through them to generate samples for the BLANC-tune dataset.
The algorithm for BLANC-tune is shown in more detail in Figure 4.
Given: |
; ; ; |
probability of masking tuning; |
min length of word to be masked ; |
number of tuning passes |
# Tune the model on the summary |
number of words in summary |
Initialize empty tuning dataset |
for in range from to : |
positions of words longer than |
Random shuffle |
until all position in are used: |
Mask words in next positions |
Add summary with masked words to |
Tune on . Result: |
# Compare inference with vs. |
Initialise , , , to zero |
for in : |
for in range from to : |
Mask each th word if |
for each position in masked tokens: |
Similar to BLANC-help, there can be several variations of the measure. The details described in the previous section for BLANC-help are now applicable here in two parts of the algorithm where we must select masked tokens: for the tuning dataset, and for the inference. Any fixed version of the measure can be reproducible, with fixed seed for randomness at the tuning. In our tuning we used the same optimizer and learning rate as was used by the open source huggingface repository Wolf et al. (2019) for training, and we found that dependency on the seed is very weak.
While BLANC-tune appears more complicated than BLANC-help, it is a promising method in that learning from a summary is separated completely from the task of understanding the document, with no concatenation required. While we use BLANC-help for the presentation of our approach in this paper, in future work we will systematically explore BLANC-tune. Our preliminary experiments showed that BLANC-tune and BLANC-help return similar values.
2.4 Extractive summaries: no-copy-pair guard
In the case of purely extractive summaries, the process of calculating BLANC scores may pair a summary with sentences from the text that have been copied into the summary. This exact sentence copying should be unfairly helpful in unmasking words in the original sentence. This effect may be reduced or completely eliminated by using a stronger underlying language model, especially for BLANC-tune. But a simpler solution is to include a simple guard rule into the measure: We may exclude any pairing of exact copy sentences from the calculation of the measure. In the process of iterating over text sentences, whenever a sentence contains its exact copy in the summary, it is skipped (or, alternative version, the copy is removed from the summary for this specific step in the process).
Throughout the paper we do not use the ”no-copy-pair” guard, except in the corner case consideration of copying random sentences from the text, as described in the next section.
3 Basic validation of BLANC measurement
As part of the validation of these new measures we performed experiments to determine how a substitution of an obviously bad summary affects the measure. One example is a summary generated by selecting random words from the text. The random words summary is generated with the same length as the original summary. Our original summaries are generated for randomly selected daily news by three different methods: by Microsoft’s abstractive UniML model Dong et al. (2019), by semi-abstractive summarization model (based on Vasilyev et al. (2019)), and by extractive LexRank model (based on Erkan and Radev (2004)). The summaries generated by these models are not flawless and vary widely in overall quality when evaluated by human labelers.
In another validation experiment, we generate a ”random sentences summary”, which is constructed from the sentences of a document. For this example, we apply BLANC-help with the ”no-copy-pair” guard introduced above. But we use the second version of the guard rule, because it is less exclusive of text sentences overall, and we also compensate for the length of the summary by replacing the copy-sentence of the summary with another sentence, rather than simply removing the copy-sentence.
BLANC-help results for both examples (in comparison to the measure of the original summaries) are shown in Figure 5.

We can see that the BLANC value for the real generated summary is almost always higher than the value for the random-sentences summary. This confirms that the measure takes into account the context as well as the informativeness of the summary to assess the quality.
Selecting only summaries with exactly three sentences, we can observe how BLANC-help deteriorates if we spoil some of the sentences of the summary. We replace one, two or all three sentences with random words, keeping the same length of the resulting randomized summary as the original summary. We also take care to run on each possible choice of replacement sentences twice, and average the resulting BLANC-help. The result is shown up in Figure 6.

4 Comparison with human evaluation scores
The BLANC measures do not require any ”gold-labeled” data: No human-written summaries nor human-annotated quality scores are needed. Theoretically, the measures should reflect how fluent, informative, and factually correct a summary is, simply because only fluent, informative, correct summaries are helpful to the underlying language model. We now turn to the question of whether the BLANC measures correlate with summary quality scores assigned by human readers.
Human scoring is fallible; a correlation with human scores should not be considered as a full validation of our measures, but rather as an independent confirmation that the measures are sensible.
For purposes unrelated to this study, we have undertaken a series of human evaluations of many generated summaries of approximately similar length. As mentioned in the previous section, the summaries were generated by Microsoft’s abstractive UniML model Dong et al. (2019), by semi-abstractive model Vasilyev et al. (2019), and by extractive LexRank model Erkan and Radev (2004). The summaries from the latter two sources were ”equalized” in length to the UniML, so that at least on average the summaries from all three generation sources would be equal, and also so that most summaries would not differ significantly in length. Altogether, we assembled 555 summary-text pairs for human scoring, with the texts taken from the CNN / Daily Mail dataset Hermann et al. (2015).
We hired 10 annotators through Odetta.ai and trained them to assess the overall quality of each summary on a 5-point scale: 0 = VERY BAD, 1 = BAD, 2 = OK, 3 = GOOD or 4 = VERY GOOD. The annotators worked independently from each other and had access to only one summary-text pair at a time. The task was performed through the online text annotation tool LightTag (lighttag.io).
The values of the correlations are illustrated in Figure 7.

The green step-function shows the value of correlation of an annotator score (with Id ranging from 1 to 10) with the averaged score of the 9 other annotators. The number of samples used for the correlation is 555 - the summaries generated by the three models. The red and blue lines show correlations of BLANC-help and rouge correspondingly with the averaged score of all 10 annotators. The rouge here is calculated using the google-research package (github.com/google-research/google-research/tree/master/rouge) as F1 value of ”rougeL” (lower blue line on the plot) and F1 value of ”rougeLsum” (upper blue line). The latter is the ’summary-level LCS’, with summaries split to sentences and using a union longest common subsequence Lin (2004)
The yellow line in the figure shows how a simplest combination of BLANC-help and ROUGE correlates with the annotators. The ”BLANC-help + ROUGE-Lsum” is literally a simple sum of BLANC-help and the ROUGE-Lsum. As usual a blending of two different models produces better results, though it is not our purpose here to fit human scores, and we do not fit the weights in the sum. (For example, using a with the weight for BLANC-help would increase the correlation with human scores by 1%).
All shown correlations have p-values of order and lower. We observe that both BLANC-help and ROUGE correlate with annotators as good as or better than about 30% of annotators.
In Figure 8 we present correlations with human scores on summaries generated for 100 typical daily news documents. The summaries were generated by the same three models; there were 300 summary-text pairs for scoring, again by 10 annotators.

Since there are no ”gold-labeled” summaries for these news documents, there is no ROUGE score in the figure.
As we see in all these examples, the human-human agreement is not impressive. We have observed from yet another evaluation dataset that if the texts and the generated summaries are challenging with very low inter-annotator agreement, the correlation of our measure with human scores is similarly diminished (with borderline p-values).
The values assigned to humans scores (0,1,2,3,4) are not a perfect translation of the human perception of the corresponding labels (”very bad”, ”bad”, ”OK”, ”good”, ”very good”). From multiple evaluations unrelated to this study we know that when an evaluation is repeated, human annotators are far more likely to substitute ”OK” and ”good” with each other than other tags. When we obtain an averaged human score, a weighting of the values with may be more inline with human perception, but the changes to results we presented here are not essential, of order 1%.
BERTScore Zhang et al. (2020), similarly to ROUGE, requires reference summaries, but is using overlaps of BERT embeddings rather than strings. In Figure 7 the BERTScore F1 would be at 0.35 - close to BLANC, BERTScore Precision at 0.16, and BERTScore Recall at impressive 0.47 (calculated using python package bert-score).
A few simple observations may serve as an evidence that our measure deals with the length of a summary more reasonably than either humans or ROUGE. In Table 1 we show the correlations with the summary length and with the compression factor, which is defined as the ratio of summary length to document text length. The length here is the number of characters.
Estimator | Correlation | L | C |
---|---|---|---|
BLANC-help | Pearson | 0.47 | 0.75 |
Spearman | 0.51 | 0.76 | |
rouge-L | Pearson | -0.27 | |
Spearman | -0.22 | ||
rouge-L-sum | Pearson | -0.23 | |
Spearman | -0.15 | ||
humans | Pearson | 0.41 | 0.31 |
Spearman | 0.41 | 0.43 |
The table is based on the same data as Figure 7. The table shows that similarly to humans, our measure is helped by longer summaries in general. But unlike humans, it is much more sensitive to a summary’s compression factor. A disregard for the compression factor by humans may be caused by the anchoring effect.
Table 2 gives similar insight for very different kind of documents - random daily news, same as were used for Figure 8.
Estimator | Correlation | L | C |
---|---|---|---|
BLANC-help | Pearson | 0.20 | 0.77 |
Spearman | 0.19 | 0.73 | |
humans | Pearson | 0.42 | 0.41 |
Spearman | 0.38 | 0.39 |
Whenever we used BLANC or human annotators for comparison of quality of summaries generated by different models of by different versions of a model, we generated summaries on average of the same length. It is clear that both humans and BLANC will estimate longer summary better, at least when it is a single score of overall summary quality. If the length of individual summary has to be excluded as a factor, the BLANC score should be normalized by the compression . A longer summary adds proportionally more help, while a longer text adds proportionally more tokens for masking.
In Table 3 we show comparison of BLANC with negated Jensen-Shannon divergence (JS) which is a no-references measure showed up as the strongest in Louis and Nenkova (2009). The JS is a mean of text-summary and summary-text Kullback-Leibler divergences. For a purely statistical measure which we would assume misses a lot of semantics, JS works surprisingly well on CNN / Daily Mail news examples. The modest performance at first row by both measures can be explained by high variety of in styles of the summaries, which affects both the human scoring and the measures. On human-only summaries JS is still better than BLANC. In order to confirm that BLANC grasps more semantics, we considered three subsets of summaries that might have less signal from pure statistics. The summaries of similar length, close to peak of the distribution, is one example; summaries with low human scores is another one. More important example is the highly compressed summaries, with the ratio of the summary length to the text length . In this case JS correlation value would be , but p-value= is too high. Following Louis and Nenkova (2009), the JS was calculated with filtering stop words and with stemming.
Selection | N | BLANC | JS |
---|---|---|---|
All | 855 | 0.22 | 0.28 |
Human | 300 | 0.34 | 0.37 |
Close length | 319 | 0.15 | 0.13 |
Bad | 141 | 0.23 | 0.21 |
Compressed | 155 | 0.18 | (p=0.15) |
Simple correlation with a consensus score of annotators is not an easy criterion for judging the usefulness of the measure. When annotators are tasked with scoring several different qualities of a summary, their final score for the overall quality should be more grounded, because more attention has been spent on the summary and the text. In Figure 9 we show values of correlations obtained from such evaluation.

The data used here are the same as the data for Figure 8: summaries generated on randomly selected daily news documents. For this illustration, however, we split our 10 annotators into a small group of 3 and an ”others” group of the remaining 7. There are 120 ways to chose the split (on the X-axis). The circle markers show human-human correlation, i.e. the correlation between the average score of the small group and the average score of the ”others” group. The plus markers show BLANC-human correlation, i.e. a correlation of the BLANC with the ”others” group of annotators. Hence we see how well the BLANC measure performs against the team of 3 annotators in correlating with the ”others”. For simplicity of the presentation, each type of correlation was sorted independently. If a correlation is unreliable (p-value ) then the marker is not shown.
We see that BLANC can be competitive to a team of three human annotators on all summary qualities, especially on the ’overall’ and fluency.
5 Conclusion
In this paper we present BLANC, a new family of objective and reproducible measures of summary quality. BLANC does not require human-written reference summaries; it is based on how helpful the summary for understanding the document.
By comparison, it is difficult to suspend disbelief when considering a method like ROUGE that does not inspect the document itself when estimating the quality of a summary. It is notable that ROUGE scores are often cited even for headline generation Ayana et al. (2016); Kiyono et al. (2017); Xu and Fung (2019); Gu et al. (2020) where it is hard to imagine that any single headline could be regarded as the best possible headline for a document.
One may argue that ROUGE requires less processing power, unless we recall that applying it requires the processing power of a human who must write the reference summary for ROUGE. In future research we will consider variations of BLANC and, for convenience, provide a public package blanc.
We thank Charlene Chambliss (Primer) for help in preparing the design of human evaluations, and Rosanne Liu (UberAI), Nina Lopatina (In-Q-Tel) and anonymous reviewers for review of the paper and valuable feedback.
References
- Ayana et al. (2016) Ayana, Shiqi Shen, Yu Zhao, Zhiyuan Liu, and Maosong Sun. 2016. Neural headline generation with sentence-wise optimization. arXiv, arXiv:1604.01904v2.
- Chen et al. (2018) Ping Chen, Fei Wu, Tong Wang, and Wei Ding. 2018. A semantic qa-based approach for text summarization evaluation. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, pages 4800–4807. AAAI Press (2018).
- Cohan and Goharian (2016) Arman Cohan and Nazli Goharian. 2016. Revisiting summarization evaluation for scientific articles. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC’16), pages 806–813. European Language Resources Association (ELRA, 2016).
- Devlin et al. (2018) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv, arXiv:1810.04805.
- Dong et al. (2019) Li Dong, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Ming Zhou Jianfeng Gao, and Hsiao-Wuen Hon. 2019. Unified language model pre-training for natural language understanding and generation. arXiv, arXiv:1905.03197.
- Elghannam and El-Shishtawy (2015) Fatma Elghannam and Tarek El-Shishtawy. 2015. Keyphrase based evaluation of automatic text summarization. arXiv, arXiv:1505.06228.
- Erkan and Radev (2004) Güneş Erkan and Dragomir R. Radev. 2004. Lexrank: Graph-based centrality as salience in text summarization. Journal of Artificial Intelligence Research, 22(1):457–479.
- Eyal et al. (2019) Matan Eyal, Tal Baumel, and Michael Elhadad. 2019. Question answering as an automatic evaluation metric for news article summarization. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1, pages 3938–3948. Association for Computational Linguistics.
- Fan et al. (2018) Lisa Fan, Dong Yu, and Lu Wang. 2018. Robust neural abstractive summarization systems and evaluation against adversarial information. arXiv, arXiv:1810.06065.
- Ganesan (2018) Kavita Ganesan. 2018. Rouge 2.0: Updated and improved measures for evaluation of summarization tasks. arXiv, arXiv:1803.01937.
- Gao et al. (2020) Yang Gao, Wei Zhao, and Steffen Eger. 2020. Supert: Towards new frontiers in unsupervised evaluation metrics for multi-document summarization. arXiv, arXiv:2005.03724.
- Gu et al. (2020) Xiaotao Gu, Yuning Mao, Jiawei Han, Jialu Liu, Hongkun Yu, You Wu, Cong Yu, Daniel Finnie, Jiaqi Zhai, and Nicholas Zukoski. 2020. Generating representative headlines for news stories. In WWW ’20: Proceedings of The Web Conference 2020, pages 1773–1784. Association for Computing Machinery, New York, 2020.
- Hardy et al. (2019) Hardy, Shashi Narayan, and Andreas Vlachos. 2019. Ehighres: Highlight-based reference-less evaluation of summarization. arXiv, arXiv:1906.01361.
- Hermann et al. (2015) Karl Moritz Hermann, Tomáš Kočiský, Edward Grefenstette, Lasse Espeholt, Will Kay, Mustafa Suleyman, and Phil Blunsom. 2015. Teaching machines to read and comprehend. In Advances in Neural Information Processing Systems 28, pages 1693–1701. Curran Associates, Inc.
- Kiyono et al. (2017) Shun Kiyono, Sho Takase, Jun Suzuki, Naoaki Okazaki, Kentaro Inui, and Masaaki Nagata. 2017. Source-side prediction for neural headline generation. arXiv, arXiv:1712.08302.
- Kryściński et al. (2019a) Wojciech Kryściński, Nitish Shirish Keskar, Bryan McCann, Caiming Xiong, and Richard Socher. 2019a. Neural text summarization: A critical evaluation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 540–551. Association for Computational Linguistics.
- Kryściński et al. (2019b) Wojciech Kryściński, Bryan McCann, Caiming Xiong, and Richard Socher. 2019b. Evaluating the factual consistency of abstractive text summarization. arXiv, arXiv:1910.12840.
- Lin (2004) Chin-Yew Lin. 2004. Rouge: A package for automatic evaluation of summaries. In Proceedings of Workshop on Text Summarization Branches Out, pages 74–81.
- Louis and Nenkova (2009) Annie Louis and Ani Nenkova. 2009. Automatically evaluating content selection in summarization without human models. In Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing, pages 306–314. Association for Computational Linguistics.
- Louis and Nenkova (2013) Annie Louis and Ani Nenkova. 2013. Automatically assessing machine summary content without a gold standard. Computational Linguistics, 39(2):267–300.
- Mao et al. (2019) Yuning Mao, Liyuan Liu, Qi Zhu, Xiang Ren, and Jiawei Han. 2019. Facet-aware evaluation for extractive text summarization. arXiv, arXiv:1908.10383.
- Ng and Abrecht (2015) Jun-Ping Ng and Viktoria Abrecht. 2015. Better summarization evaluation with word embeddings for rouge. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 1925–1930. Association for Computational Linguistics.
- Scialom et al. (2019) Thomas Scialom, Sylvain Lamprier, Benjamin Piwowarski, and Jacopo Staiano. 2019. Answers unite! unsupervised metrics for reinforced summarization models. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 3246–3256, Hong Kong, China. Association for Computational Linguistics.
- Shao et al. (2017) Liqun Shao, Hao Zhang, Ming Jia, and Jie Wang. 2017. Efficient and effective single-document summarizations and a word-embedding measurement of quality. arXiv, arXiv:1710.00284.
- Taylor (1953) Wilson L Taylor. 1953. Cloze procedure: A new tool for measuring readability. Journalism Bulletin, 30(4):415–433.
- Vasilyev et al. (2019) Oleg Vasilyev, Tom Grek, and John Bohannon. 2019. Headline generation: Learning from decomposable document titles. arXiv, arXiv:1904.08455v3.
- Wolf et al. (2019) Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, and Jamie Brew. 2019. Huggingface’s transformers: State-of-the-art natural language processing. arXiv, arXiv:1910.03771.
- Xenouleas et al. (2019) Stratos Xenouleas, Prodromos Malakasiotis, Marianna Apidianaki, and Ion Androutsopoulos. 2019. Sumqe: a bert-based summary quality estimation model. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 6005–6011, Hong Kong, China. Association for Computational Linguistics.
- Xu and Fung (2019) Peng Xu and Pascale Fung. 2019. A novel repetition normalized adversarial reward for headline generation. In ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 7325–7329, Brighton, United Kingdom. IEEE.
- Zhang et al. (2020) Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian Q. Weinberger, and Yoav Artzi. 2020. Bertscore: Evaluating text generation with bert. arXiv, arXiv:1904.09675v3.
- Zhao et al. (2019) Wei Zhao, Maxime Peyrard, Fei Liu, Yang Gao, Christian M. Meyer, and Steffen Eger. 2019. Moverscore: Text generation evaluating with contextualized embeddings and earth mover distance. arXiv, arXiv:1909.02622.
- Ziegler et al. (2020) Daniel M. Ziegler, Nisan Stiennon, Jeffrey Wu, Tom B. Brown, Alec Radford, Dario Amodei, Paul Christiano, and Geoffrey Irving. 2020. Fine-tuning language models from human preferences. arXiv, arXiv:1909.08593v2.