This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Corpora Evaluation and System Bias Detection
in Multi-document Summarization

Alvin Dey  Tanya Chowdhury{}^{\diamondsuit^{\dagger}}11footnotemark: 1 Yash Kumar Atri11footnotemark: 1  Tanmoy Chakraborty
University of Massachusetts, Amherst, MA, USA
IIIT-Delhi, New Delhi, India
[email protected]
{alvin18066,yashk,tanmoy}@iiitd.ac.in
Equal contribution; listed alphabetically.
Abstract

Multi-document summarization (MDS) is the task of reflecting key points from any set of documents into a concise text paragraph. In the past, it has been used to aggregate news, tweets, product reviews, etc. from various sources. Owing to no standard definition of the task, we encounter a plethora of datasets with varying levels of overlap and conflict between participating documents. There is also no standard regarding what constitutes summary information in MDS. Adding to the challenge is the fact that new systems report results on a set of chosen datasets, which might not correlate with their performance on the other datasets. In this paper, we study this heterogeneous task with the help of a few widely used MDS corpora and a suite of state-of-the-art models. We make an attempt to quantify the quality of summarization corpus and prescribe a list of points to consider while proposing a new MDS corpus. Next, we analyze the reason behind the absence of an MDS system which achieves superior performance across all corpora. We then observe the extent to which system metrics are influenced, and bias is propagated due to corpus properties. The scripts to reproduce the experiments in this work are available at https://github.com/LCS2-IIITD/summarization_bias.git.

1 Introduction

Multi-document summarization (MDS) deals with compressing more than one document into a textual summary. It has a wide range of applications – gaining insights from tweets related to similar hashtags, understanding product features amongst e-commerce reviews, summarizing live blogs related to an ongoing match, etc. Most studies on MDS were performed during the DUC111https://duc.nist.gov/ and TAC222http://tac.nist.gov challenges starting in the early 2000s. Each version of the challenges released a new dataset. Most of the MDS systems submitted to these challenges were unsupervised and extractive in nature. Gradually, the data released in these challenges became the de facto for MDS. These datasets were manually curated and had less than a hundred instances each. The recent development of deep neural architecture has led to a significant increase in the number of supervised document summarization systems. Large labeled corpora which are mostly crowd-sourced have been introduced to meet the training requirements of the supervised systems. However, the crowd-sourced corpora widely differ in quality based on factors like genre, size of the community, presence of moderation in the community, etc. This is further aggravated by the complexity of the task, the hardness of accumulating labeled data, or more so in the definition of what constitutes a multi-document summary.

Recently, a few large datasets for MDS have been introduced (Fabbri et al., 2019; Chowdhury and Chakraborty, 2019a). However, there has been no study to measure the relative complexity of these datasets. We observe that existing MDS systems behave differently on different corpora. For example, a system achieving state-of-the-art performance on one corpus fails to achieve reasonable performance on another. Although the ROUGE points of MDS systems are increasing day-by-day, manual inspection reveals an increased presence of bias in generated summaries. New systems are being introduced and evaluated on a few selected corpora, leading to difficulty in understanding whether the bias is introduced by the system or it is present in the corpus used for training.

Our research questions are as follows:
Q1. How should one model the quality of a MDS corpus as a function of its intrinsic properties?
Q2. Why do the ROUGE-based ranks of different MDS systems differ across different corpora? How should an MDS system which intends to achieve high ROUGE scores across all corpora, look like?
Q3. Why do systems show bias on different metrics, and which other system and corpus attributes are the reason behind it?
Q4. Is the task of MDS almost solved, or is there still scope for improvement?
We study five MDS corpora – DUC (DUC, 2002), TAC (TAC, 2008), Opinosis (Ganesan et al., 2010), Multinews (Fabbri et al., 2019), and CQASumm (Chowdhury and Chakraborty, 2019b). We consider eight popular summarization systems – LexRank (Erkan and Radev, 2004), TextRank (Mihalcea and Tarau, 2004), MMR (Carbinell and Goldstein, 2017), ICSISumm (Gillick et al., 2008), PG (See et al., 2017), PG-MMR (Lebanoff et al., 2018), Hi-Map (Fabbri et al., 2019), and CopyTransformer (Gehrmann et al., 2018).

Our major contributions are four-fold:
\bullet We propose a suite of metrics to model the quality of an MDS corpus in terms of – Abstractness, Inter Document Similarity (IDS), Redundancy, Pyramid Score, Layout Bias and Inverse-Pyramid Score.
\bullet We develop an interactive web portal for imminent corpora to be uploaded and evaluated based on our proposed metrics.
\bullet We explore different biases that the MDS systems exhibit over different corpora and provide insight into properties that a universal MDS system should display to achieve reasonable performance on all types of corpora.
\bullet We look into metrics to capture bias shown by MDS systems and explore the extent to which corpus properties influence them.

To the best of our knowledge, the current study is the first of its kind.

2 Background and Proposed Metrics

Throughout the paper, we use the term candidate documents for the documents participating in summarization, and the term reference to indicate the ground-truth summary.

Oracle summary is the extractive set of sentences selected from the candidate documents, exhibiting maximum ROUGE-N score w.r.t. the reference summary. It is an NP-hard problem Hirao et al. (2017), and approximate solutions can be found greedily or using ILP solvers.

Here, we briefly introduce a suite of corpus and system metrics proposed by us to better understand the MDS task. These metrics are further explained in detail in Supplementary.

2.1 Corpus Metrics

\bullet Abstractness: It is defined as the percentage of non-overlapping higher order nn-grams between the reference summary and candidate documents. A high score highlights the presence of more distinctive phrases in reference summary. The intuition behind quantifying the number of new words is to sync with the basic human nature of paraphrasing while summarizing.
\bullet Inter Document Similarity (IDS): It is an indicator of the degree of overlap between candidate documents. Inspired by the theoretical model of relevance (Peyrard, 2019a), we calculate IDS of a set of documents as follows:

IDS(Di)=DjSRelevance(Dj,Di)|S|IDS(D_{i})=\frac{\sum_{D_{j}\in S}Relevance(D_{j},D_{i})}{|S|}\vspace{-1mm} (1)

where DiD_{i} is the ithi^{th} candidate document, and SS is the set of all documents other than DiD_{i}. Here, RelevanceRelevance(.,.) is defined as:

Relevance(A,B)=ωiPA(ωi).log(PB(ωi))Relevance(A,B)=\sum_{\omega_{i}}P_{A}(\omega_{i}).\log(P_{B}(\omega_{i}))\vspace{-1mm} (2)

where PA(ωi)P_{A}(\omega_{i}) represents the probability distribution of the ithi^{th} semantic unit333An atomic piece of information in document AA. The further this score is from 0, the lesser inter document overlap there is in terms of semantic unit distribution. As shown in Equation 1, the numerator calculates relevance which can be interpreted as the average surprise of observing one distribution while expecting another. This score is small if the distributions are similar i.e., PAP_{A} \approx PBP_{B} from Equation 2.

\bullet Pyramid Score: We propose the metric Corpus Pyramid score to measure how well important information across documents is represented in the ground truth. As introduced by Nenkova and Passonneau (2004), Pyramid score is a metric to evaluate system summaries w.r.t. the pool of ground-truth summaries. We instead use this metric to quantitatively analyze the ground-truth summary w.r.t. candidate documents. The entire information set is split into Summarization Content Units (SCUs444They are subsentential units based on semantic meaning), and each SCU is assigned a weight based on the number of times it occurs in the text. A pyramid of SCUs is constructed with an SCU’s weight, denoting its level, and a score is assigned to a text, based on the number of SCUs it contains. Pyramid score is defined as the ratio of a reference summary score and an optimal summary score. Higher values indicate that the reference summary covers the SCUs at the top of the pyramid better. SCUs present at the top are the ones occurring in most articles and thus can be deemed as important.

\bullet Redundancy: The amount of information in a text can be measured as the negative of Shannon’s entropy (HH) (Peyrard, 2019a).

H(D)=ωiPD(ωi).log(PD(ωi))H(D)=-\sum_{\omega_{i}}P_{D}(\omega_{i}).\log(P_{D}(\omega_{i}))\vspace{-1mm} (3)

where PDP_{D} represents the probability distribution of documents DD, and ωi\omega_{i} represents the ithi^{th} semantic unit3 in the distribution. H(D)H(D) would be maximized for a uniform probability distribution when each semantic unit is present only once. The farther this score is from 0, the better a document is distributed over its semantic units in the distribution, hence lesser the redundancy. As evident from Equation 5, redundancy is maximized if all semantics units have equal distribution i.e., P(ωi)P(\omega_{i}) = P(ωj)P(\omega_{j}). The idea behind using redundancy is to quantify how well individual documents cover sub-topics, which might not be the core content but important nonetheless. Thus

Redundancy(D)=HmaxH(D)Redundancy(D)=H_{max}-H(D)\vspace{-1mm} (4)

Since HmaxH_{max} is constant, we obtain

Redundancy(D)=ωiPD(ωi).log(PD(ωi))Redundancy(D)=\sum_{\omega_{i}}P_{D}(\omega_{i}).\log(P_{D}(\omega_{i}))\vspace{-1mm} (5)

\bullet Layout Bias: We define Layout Bias across a document as the degree of change in importance w.r.t. the ground-truth over the course of candidate documents. We divide the document into kk segments, calculate the importance of each segment w.r.t. the ground-truth by a similarity score, and average over the sentences in the segment. Positional importance of DjD_{j}, the jthj^{th} sentence in the document is denoted by:

PositionalImportance(Dj)=max1insim(Dj,Ri)PositionalImportance(D_{j})=\max_{1\leq{i}\leq{n}}sim(\overrightarrow{D_{j}},\overrightarrow{R_{i}})\vspace{-1mm} (6)

where, Ri\overrightarrow{R_{i}} is the vector representation of the ithi^{th} sentence in the reference, simsim is a similarity metric between two sentences, and nn is the total number of sentences in the reference summary.
A lower shift indicates that while generating reference summaries, all segments have been given similar importance within any 3-fold segmented article.
\bullet Inverse-Pyramid Score (Inv Pyr): We propose Inverse-Pyramid score to quantify the bias that a reference summary exhibits w.r.t. its set of candidate documents. It measures the importance given to each document in the candidate set by the reference summary as:

InvPyr(D,S)=Varj(DjSu)InvPyr(D,S)=Var_{j}\big{(}{D_{j}}\cap{S_{u}}\big{)}\vspace{-1mm} (7)

Here, DD and SS are the set of candidate documents for MDS and their summary respectively, VarVar is the variance, DjD_{j} and SuS_{u} are the sets of SCUs4 in the jthj^{th} document of the candidate set and the reference summary respectively.

Higher Inv Pyr scores suggest the difference in importance given to each document while generating the summary is higher. As evident from Equation 7, Variance across the similarities is high if the similarity scores across the document-summary pairs are uneven.

Refer to caption
Figure 1: Heatmap depicting the corpus metric: Inter document similarity. We explain with a single instance of (a) DUC-2004, (b) DUC-2003, (c) TAC-2008, and (d) CQASumm, and highlight inter-document overlap.
Refer to caption
Figure 2: (a) Layout Bias across datasets, highlighting cumulative cosine similarity (importance) values (y-axis) between segments (first, second and third) of candidate documents and the reference summary. (b) Change in layout importance across systems over source segments when divided in three uniform segments. (c) Change in layout importance across systems when candidate documents are internally shuffled and divided into three uniform segments.

2.2 System Metrics

\bullet ROUGE (Lin, 2004) is a metric which computes the nn-gram overlap recall value for the generated summary w.r.t. the reference summary.
\bullet F1 Score with Oracle Summaries: Oracle summaries reflect the extractive selection of sentences that achieve the highest ROUGE score over the candidate documents given a reference summary. Similar to ROUGE-1, this metric also combines both precision and recall between the oracle and system summaries to calculate F1 Score. It is a better indicator of the presence of non-essential nn-grams than ROUGE as it also takes precision into account.
\bullet System Abstractness: Analogous to corpus abstractness, we compute the percentage of novel higher order nn-grams in the generated summary w.r.t. the candidate documents. System abstractness is calculated using

Coverage(D,S)=i1..n(DSi)Cn(S)Coverage(D,S)=\frac{\sum_{i\in{1..n}}(D\cap S_{i})}{C_{n}(S)}

where DD represents the set of n-grams in the candidate document, and SS represents the set of n-grams in the ithi^{th} system summary.
The denominator denotes the total count of n-grams in a system summary. Finally, the values of all articles is normalized to get the score for the system
\bullet Layout Bias: We propose this metric to capture which sections of the candidate documents comprise a majority of the information in the generated summary. For neural abstractive systems, we concatenate candidate documents to form one large document and feed it to the neural model. We study two variations of this metric – The first variation involves segmenting this large document into kk parts and then computing the similarity of nn-gram tokens of system summaries w.r.t. the candidate document segment. The second variation includes shuffling the candidate documents before concatenating and then computing the nn-gram similarity with the generated summary.
\bullet Inter Document Distribution (IDD): We propose this metric to quantify the extent of contribution of each candidate document to form the generated summary. The relevance for system summaries is calculated by,

Relevance(A,B)=ωiPA(ωi).log(PB(ωi))Relevance(A,B)=\sum_{\omega_{i}}P_{A}(\omega_{i}).\log(P_{B}(\omega_{i}))

where PAP_{A} represents the probability distribution of system summary SS, and ωi\omega_{i} represents the ithi^{th} semantic unit in the distribution.

IDD(Di)=DjSRelevance(Dj,Di)Cardinality(S)IDD(D_{i})=\frac{\sum_{D_{j}\in S}Relevance(D_{j},D_{i})}{Cardinality(S)}

\bullet Redundancy: It measures the degree to which system summaries can cover the distribution across semantic units generated from the candidate documents. Redundancy for candidate documents is given by Eq.,

Redundancy(D)=ωiSD(ωi).log(SD(ωi))Redundancy(D)=\sum_{\omega_{i}}S_{D}(\omega_{i}).\log(S_{D}(\omega_{i}))

where SDS_{D} represents the probability distribution of a system summary DD. ωi\omega_{i} represents the ithi^{th} semantic unit in the distribution.

3 Experimental Setup

3.1 MDS Corpora

\bullet DUC (DUC, 2002) is a news dataset built using newswire/paper documents. The 2003 (DUC-2003) and 2004 (DUC-2004) versions comprise 3030 and 5050 topics respectively with each topic having 4 manually curated reference summaries.
\bullet TAC (TAC, 2008) is built from the AQUIANT-2 collection of newswire articles where NIST assessors select 4848 and 4444 topics for the 2008 and 2010 versions, respectively. Each topic consists of 4 summaries.
\bullet Opinosis (Ganesan et al., 2010) is an accumulation of user reviews collected from various sources like TripAdvisor, Edmunds.com and Amazon. There are 5151 topics, with each topic having approximately 44 human-written summaries.
\bullet CQASumm (Chowdhury and Chakraborty, 2019b) is a community question answering dataset, consisting of 100,000100,000 threads from the Yahoo! Answers L6 dataset. It treats each answer under a thread as a separate document and the best answer as a reference summary.
\bullet Multinews (Fabbri et al., 2019) is a news dataset comprising news articles and human-written summaries from newser.com. It has 56,21656,216 topics, with summaries of 260260 words on average written by professional editors.

3.2 MDS Systems

To identify bias in system-generated summaries, we study a few non-neural extractive and neural abstractive summarization systems, which are extensively used for multi-document summarization.
\bullet LexRank (Erkan and Radev, 2004) is a graph based algorithm that computes the importance of a sentence using the concept of eigen vector centrality in a graphical representation of text.
\bullet TextRank (Mihalcea and Tarau, 2004) runs a modified version of PageRank (Brin and Page, 1998) on a weighted graph, consisting of nodes as sentences and edges as similarities between sentences.
\bullet Maximal Marginal Relevance (MMR) (Carbinell and Goldstein, 2017) is an extractive summarization system that ranks sentences based on higher relevance while considering the novelty of the sentence to reduce redundancy.
\bullet ICSISumm (Gillick et al., 2008) optimizes the summary coverage by adopting a linear optimization framework. It finds a globally optimal summary using the most important concepts covered in the document.
\bullet Pointer Generator (PG) network (See et al., 2017) is a sequence-to-sequence summarization model which allows both copying words from the source by pointing or generating words from a fixed vocabulary.
\bullet Pointer Generator-MMR: PG-MMR (Lebanoff et al., 2018) uses MMR along with PG for better coverage and redundancy mitigation.
\bullet Hi-Map: Hierarchical MMR-Attention PG model (Fabbri et al., 2019) extends the work of PG and MMR. MMR scores are calculated at word level and incorporated in the attention weights for a better summary generation.
\bullet Bottom-up Abstractive Summarization (CopyTransformer) (Gehrmann et al., 2018) uses transformer parameters proposed by Vaswani et al. (2017); but one of the attention heads chosen randomly acts as a copy distribution.

Table 1: Values of corpus metrics: Abstractness, Redundancy (Red), Inter Document Similarity (IDS), Pyramid Score (Pyr) and Inverse-Pyramid Score (Inv).
Dataset Metric
Abstractness Red IDS Pyr Inv
1-gram 2-gram 3-gram
DUC 11.5 54.66 79.29 -0.21 -6.6 0.35 2.64
Opinosis 11.5 50.36 76.31 -0.02 -5.53 0.26 2.8
Multinews 32.28 67.53 80.45 -0.8 -1.03 0.4 3.8
CQASumm 41.41 80.72 88.79 -0.22 -9.16 0.05 5.2
TAC 9.91 50.26 76.17 -0.19 -4.43 0.32 2.9
Table 2: Various metrics (Met) showing ROUGE Scores (ROUGE-1, ROUGE-2), F1 Score (F1) between candidate documents and oracle summaries, Abstractness (Abs.) of abstractive systems, Redundancy (Red.) in system generated summaries, Inter Document Distribution (IDD) and Inter Document Distribution Variance (IDDV) of system summaries in dataset DUC, TAC, Opin (Opinosis), Multin (Multinews and CQAS (CQASumm).
System Met. Dataset
DUC TAC Opin Multin CQAS
Lex -Rank R1 35.56 33.1 33.41 38.27 32.22
R2 7.87 7.5 9.61 12.7 5.84
F1 31.34 31.51 31.05 41.01 49.71
Red. -0.136 -0.104 -0.278 -0.29 -0.364
IDD -3.377 -1.87 -3.526 -2.53 -2.17
IDDV 0.239 1.62 0.221 0.242 1.232
Text -Rank R1 33.16 44.98 26.97 38.44 28.94
R2 6.13 9.28 6.99 13.1 5.65
F1 40.8 29.69 31 38.44 46.3
Red. -0.25 -1.553 -0.342 -0.208 -0.247
IDD -0.196 -5.97 -2.745 -1.879 -2.137
IDDV 0.799 1.48 0.025 0.146 0.744
MMR R1 30.14 30.54 30.24 38.77 29.33
R2 4.55 4.04 7.67 11.98 4.99
F1 30.57 28.3 31.8 42.07 45.48
Red. -0.266 -0.068 -0.255 -0.17 -0.288
IDD -2.689 -2.135 -3.213 -1.83 -2.059
IDDV 1.873 0.231 0.222 0.157 0.126
ICSI -Summ R1 37.31 28.09 27.63 37.2 28.99
R2 9.36 3.78 5.32 13.5 4.24
F1 24.27 27.82 29.83 44.71 50.98
Red. -0.327 -0.283 -0.328 -0.31 -0.269
IDD -3.357 -1.903 -3.244 -3.14 -2.466
IDDV 0.694 0.403 1.134 0.239 0.242
PG R1 31.43 31.44 19.65 41.85 31.09
R2 6.03 6.4 1.29 12.91 5.52
F1 23.08 26.32 16.08 43.89 21.85
Abs. 0.017 0.01 0.04 0.28 0.065
Red. -0.16 -0.2542 -0.188 -0.28 -0.12
IDD -2.1 -1.93 -2.1 -2.103 -0.5
IDDV 0.248 0.398 0.168 0.391 0.391
PG -MMR R1 36.42 40.44 19.8 40.55 36.54
R2 9.36 14.93 1.34 12.36 6.67
F1 24.3 26.9 16.39 43.93 21.72
Abs. 0.019 0.02 0.04 0.275 0.069
Red. -0.17 -0.26 -0.172 -0.29 -0.142
IDD -2.4 -1.87 -1.9 -1.98 -0.72
IDDV 0.441 0.274 0.192 0.249 0.318
Trans. R1 28.54 31.54 20.46 43.57 30.12
R2 6.38 5.9 1.41 14.03 4.36
F1 15.72 17.82 16.38 44.54 21.35
Red. -0.1771 -0.17 -0.189 -0.18 -0.273
Abs. 0.09 0.09 0.049 0.319 0.092
IDD -1.9148 -1.8677 -1.589 -1.89 -2.239
IDDV 0.138 0.172 0.249 0.126 1.184
Hi -Map R1 35.78 29.31 18.02 43.47 31.41
R2 8.9 4.61 1.46 14.89 4.69
F1 25.89 24.3 20.36 42.55 19.84
Abs. 0.14 0.147 0.08 0.267 0.07
Red. -0.1722 -0.2002 -0.16 -0.23 -0.26
IDD -1.6201 -1.652 -1.8 -1.788 -2.223
IDDV 0.185 0.155 0.209 0.209 0.448
Highest R1 94.01 94.07 44.53 79.94 64.45
R2 49.85 50.17 5.73 42.41 18.38
Refer to caption
Figure 3: (a) Abstractness across datasets. (b) Redundancy, Pyramid Score and Inverse-Pyramid Score (Inv Pyr scaled down by a factor of 10 for better visualization with other metrics) across datasets. (c) Inter Document Similarity (IDS) across datasets.
Refer to caption
Figure 4: (a) Level of abstractness of systems w.r.t. candidate documents and the system generated summaries. (b) F1 Score of various systems between oracle summaries and system-generated summaries. (c) ROUGE scores of various system summaries on the left axis and maximum ROUGE score over a dataset on the right axis.

4 Inferences from Corpus Metrics

\bullet News derived corpora show a strong layout bias where significant reference information is contained in the introductory sentences of the candidate documents (Fig. 2).
\bullet Different MDS corpora vary in compression factors with DUC at 56.5556.55, TAC at 54.6854.68, Multinews at 8.188.18 and CQASumm at 5.655.65. A high compression score indicates an attempt to pack candidate documents to a shorter reference summary.
\bullet There has been a shift in the size and abstractness of reference summaries in MDS corpora over time – while DUC and TAC were small in size and mostly extractive (11%11\% novel unigrams); crowd-sourced corpora like CQASumm are large enough to train neural models and highly abstractive (41.4%41.4\% novel unigrams).
\bullet Candidate documents in Opinosis, TAC and DUC feature a high degree of redundant information as compared to Opinosis and CQASumm, with instances of the former revolving around a single key entity while that of the latter tending to show more topical versatility.
\bullet MDS corpora present a variation in inter-document content overlap as well: while Multinews shows the highest degree of overlap, CQASumm shows the least and the rest of the corpora show moderate overlap (see Fig. 1).
\bullet Pyramid Score, the metric which evaluates if the important and redundant SCUs4 from the candidate documents have been elected to be part of the reference summary, shows considerably positive values for DUC, TAC and Multinews as compared to crowdsourced corpora like CQASumm (Fig. 3.b).
\bullet Inverse-Pyramid Score, the metric which evaluates how well SCUs4 of the reference summary are distributed amongst candidate documents, also shows better performance on human-annotated corpora compared to crowd-sourced ones (Fig. 3(b)).
\bullet A comparison amongst corpus metrics presents a strong positive correlation between IDS and Pyramid Score (Pearson’s ρ\rho = 0.8296) and a strong negative correlation between the metrics of Redundancy and IDS (Pearson’s ρ\rho = -0.8454).

Refer to caption
Figure 5: Redundancy of various systems across DUC, TAC, Opinosis, Multinews and CQASumm.

5 Inferences from System Metrics

\bullet MDS systems under consideration are ranked differently in terms of ROUGE on different corpora; leading to a dilemma whether to declare a system superior to others without testing on all types of datasets (Fig. 4(c)) and Table 2).
\bullet MDS systems under consideration outperform abstractive summarization systems by up to 10%10\% on ROUGE-1 and up to 30%30\% on F1 Scores, showing contradictory behavior in comparison to single-document summarization systems where state-of-the-art abstractive systems are known to outperform the former (Figs. 4(b)-(c)).
\bullet The best summarization system on each corpus obtains a score 39.6%,47.8%,75.02%,54.5%,49.9%39.6\%,47.8\%,75.02\%,54.5\%,49.9\% of the oracle upper bound on DUC, TAC, Opinosis, Multinews and CQASumm respectively, indicating that summarization on Opinosis and Multinews is a partially solved problem, while DUC, TAC and CQASumm exhibit considerable scope for improvement (Fig. 4(c)).
\bullet Hi-Map and CopyTransformer generate more abstract summaries (17.5%17.5\% and 16%16\% novel unigrams respectively) in comparison to PG and PG-MMR (Fig. 4(a)).
\bullet Averaging over systems and comparing corpora, we notice that Multinews and CQASumm achieve the highest abstractness (27%27\% and 7%7\% respectively), which might be a result of these two corpora having the most abstract reference summaries (Fig. 4(a) and (Table 2)).
\bullet Abstractive systems exhibit a 55%55\% shift in importance between the first and the second segments of generated summaries, whereas extractive systems show an average shift of only 40%40\%, implying that abstractive systems have a stronger tendency to display layout bias (Fig. 2(b) and Fig. 2(c)).
\bullet While DUC, TAC and Opinosis summaries generated from PG trained models exhibit lower novel unigrams formation, the same for CopyTransformer and Hi-Map on DUC, TAC and Opinosis shows a higher unigram formation on average (Fig. 4(a)).
\bullet In terms of Inter Document Distribution, LexRank summary for TAC and CQASumm shows more variance across documents compared to DUC, Opinosis and Multinews. TextRank summary on DUC, TAC and CQASumm, MMR summary on DUC, and Hi-Map summary on CQASumm show higher variances as well. Systems such as PG, PG-MMR and CopyTransformer show minimal deviation in the document participation across corpora (Table 2).
\bullet In terms of Topic Coverage, extractive systems show better coverage than abstractive systems (Table 2), which might be a result of extractive systems being based on sentence similarity algorithms which find important sentences, reduce redundancy and increase the spread of information from different segments of the candidate document. (Fig. 5).

6 Discussion on Research Questions

Q1. How should one model the quality of an MDS corpus as a function of its intrinsic metrics? What guidelines should be followed to propose MDS corpora for enabling a fair comparison with existing datasets? The quality of an MDS corpus is a function of two independent variables: the quality of the candidate documents and the quality of the reference summary. Our findings suggest that a framework for future MDS datasets should provide scores measuring their standing w.r.t. both the above factors. The former is usually crowd-source dependent, while the latter is usually annotator dependent. While Inter Document Similarity, Redundancy, Layout Bias and Inverse-Pyramid Score are indicators of the properties of the candidate document, metrics such as Abstractness of the reference summary and Pyramid Score are ground-truth properties. We divide the above metrics into two categories: objective and subjective. While all these metrics should be reported by imminent corpora proposers to enable comparisons with existing corpora and systems, we feel that the objective metrics average Pyramid Score and Inverse-Pyramid Score must be reported as they are strong indicators of generic corpus quality. Other subjective metrics such as IDS, Redundancy, Abstractness etc. can be modeled to optimize task-based requirements.

Q2. Why do the ROUGE-based ranks of different MDS systems differ across corpora? How should an MDS system which is to achieve reasonably good ROUGE score on all corpora look like? From Table 2 within studied systems, in terms of ROUGE-1, ICSISumm achieves the best score on DUC, TextRank on TAC, LexRank on Opinosis, CopyTransformer on Multinews and LexRank on CQASumm. Hence as of today, no summarization system strictly outperforms others on every corpus. We also see that CopyTransformer which achieves state-of-the-art performance on Multinews achieves 10 points less than the best system on DUC. Similarly, LexRank, the state-of-the-art performer on CQASumm, achieves almost 12 points less than the best system on TAC. Therefore, a system that performs reasonably well across all corpora, is also missing. This is because different corpora are high on various bias metrics, and summarization systems designed for a particular corpus take advantage and even aggravate these biases. For example, summarization systems proposed on news based corpora are known to feed only the first few hundred tokens to neural models, thus taking advantage of the layout bias. Feeding entire documents to these networks have shown relatively lower performance. Systems such as LexRank are known to perform well on candidate documents with high inter-document similarity (e.g., Opinosis). Solving the summarization problem for an unbiased corpus is a harder problem, and for a system to be able to perform reasonably well on any test set, it should be optimized to work on such corpora.

Table 3: Pearson correlation between corpus and system with column 4 (First) between Abstractness of corpora and system, column 5 (Second) between Abstractness of corpora and ROUGE-1 score of systems across datasets and column 6 (Third) showing Layout Bias correlation between system and corpora.
Metric
Layout correlation
System Abs. corr R-1 corr First Second Third
LexRank - 0.08 0.88 0.06 0.96
TextRank - -0.24 0.91 0.76 0.97
MMR - 0.32 0.86 0.09 0.97
ICSISumm - 0.11 0.39 0.53 0.72
PG 0.57 0.65 0.80 -0.80 -0.98
PG-MMR 0.57 0.33 0.84 -0.69 -0.91
CopyTrans. 0.47 0.50 0.84 -0.31 -0.79
Hi - Map 0.11 0.45 0.74 -0.11 -0.46

Q3. Why do systems show bias on different metrics, and which other system and corpus attributes are the reason behind it? We begin by studying how abstractness of generated summaries is related to the abstractness of corpora the system is trained on. For this, we calculate the Pearson correlation coefficient between the abstractness of generated summaries and references across different datasets. From Table 3, we infer that PG, PG-MMR and CopyTransformer show a positive correlation which implies that they are likely to generate more abstract summaries if the datasets on which they are trained have more abstract references. Lastly, we infer how Layout Bias in system-generated summaries is dependent on the layout bias of reference summaries. The last three highlighted columns of Table 3 infer that the abstractive systems such as PG, PG-MMR, Hi-Map and CopyTransformer show a high negative correlation for the end segments while maintaining a strongly positive one with the starting segment. On the other hand, extractive systems such as LexRank, TextRank, MMR and ICSISumm maintain a strongly positive correlation throughout the segments. On shuffling the source segments internally, we observe that extractive systems tend to retain their correlation with corpora while abstractive systems show no correlation at all (Fig. 2), proving that in supervised systems, the layout bias in system summaries propagates from the layout bias present in corpora.

Q4. Is the task of MDS almost solved, or there is still plenty of scope remaining for improvement? In the previous sections, we computed the oracle extractive upper bound summary using greedy approaches to find the summary that obtains the highest ROUGE score given the candidate documents and references. We observe that the best summarization system on each corpus today obtains a score which is 39.6%39.6\% of the extractive oracle upper bound on DUC, 47.8%47.8\% on TAC, 75.02%75.02\% on Opinosis, 54.5%54.5\% on Multinews and 49.9%49.9\% on CQASumm. This shows that there is enough scope for MDS systems to achieve double the ROUGE scores obtained by the best system to date on each corpus except Opinosis. Therefore, we believe that the task of MDS is only partially solved and considerable efforts need to be devoted to improving the systems.

7 Related Work

Previous attempts to evaluate the quality of the benchmark summarization corpora are few in number and mostly from the time when corpora were manually accumulated. Hirao et al. (2004) primarily used the intrinsic metrics of precision and recall to evaluate corpus quality. In addition, the authors proposed an extrinsic metric, called ‘Pseudo Question Answering’. This metric evaluates whether a summary has an answer to a question that is otherwise answerable by reading the documents or not. Although effective, the cost of such an evaluation is enormous and is not scalable to modern day corpora sizes. For such corpora where multiple references are available, Benikova et al. (2016) used an inter-annotator agreement to model the quality of the corpora. They also used non-redundancy, focus, structure, referential clarity, readability, coherence, length, grammaticality, spelling, layout, and overall quality as quantitative features for an MDS corpus. Recently, Chowdhury et al. (2020) proposed an MDS system that used the baseline PG model along with Hierarchical structural attention to take into account long-term dependencies for superior results compared to baseline models.

There have been a series of very recent studies that look into how to strengthen the definition and discover system biases in single-document summarization. Very recently, Jung et al. (2019) studied how position, diversity and importance are significant metrics in analyzing the toughness of single-document summarization corpora. Another recent work (Kryscinski et al., 2019) extensively studied the Layout Bias in news datasets that most single-document summarization systems seem to exploit. Two seminal works, namely Peyrard (2019a) and Peyrard (2019b), exploited the theoretical complexity of summarization on the ground of importance, analyzing in-depth what makes for a good summary. Peyrard (2019a) mathematically modeled the previously intuitive concepts of Redundancy, Relevance and Informativeness to define importance in single-document summarization. Grusky et al. (2018) proposed a new single-document summarization corpus and quantified how it compares to other datasets in terms of diversity and difficulty of the data. They introduced metrics such as extractive fragment density and extractive fragment coverage to plot the quality of SDS corpus. To the best of our knowledge, no comparative work exists for either corpora or systems in MDS, and the current paper is the first in this direction.

8 Conclusion

In this paper, we aimed to study the heterogeneous task of multi-document summarization. We analyzed interactions between widely used corpora and several state-of-the-art systems to arrive at a line of conclusions. We defined MDS as a mapping from a set of non-independent candidate documents to a synopsis that covers important and redundant content present in the source. We proposed intrinsic metrics to model the quality of an MDS corpus and introduced a framework for future researches to consider while proposing a new corpus.

We analyzed how ROUGE ranks of different systems vary differently on different corpora and described what a system that achieves reasonable performance on all corpora would look like. We evaluated how different systems exhibit bias and how their behavior is influenced by corpus properties. We also commented on the future scope for the task of MDS.

Future directions to take forward this work would include a causal analysis of how corpus bias is responsible for bias in model prediction across different corpora and systems. This might bring forward measures to de-bias NLP algorithms with/without de-biasing the corpora.

Acknowledgments

The work was partially supported by the Ramanujan Fellowship and DST (ECR/2017/00l691). T. Chakraborty would like to acknowledge the support of the Infosys Center for AI, IIIT-Delhi.

References

  • Benikova et al. (2016) Darina Benikova, Margot Mieskes, Christian M Meyer, and Iryna Gurevych. 2016. Bridging the gap between extractive and abstractive summaries: Creation and evaluation of coherent extracts from heterogeneous sources. In Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, pages 1039–1050.
  • Brin and Page (1998) Sergey Brin and Lawrence Page. 1998. The anatomy of a large-scale hypertextual web search engine. In Proceedings of the Seventh International Conference on World Wide Web 7, pages 107–117.
  • Carbinell and Goldstein (2017) Jaime Carbinell and Jade Goldstein. 2017. The use of mmr, diversity-based reranking for reordering documents and producing summaries. SIGIR Forum, 51(2):209–210.
  • Chowdhury and Chakraborty (2019a) Tanya Chowdhury and Tanmoy Chakraborty. 2019a. Cqasumm: Building references for community question answering summarization corpora. Association for Computing Machinery.
  • Chowdhury and Chakraborty (2019b) Tanya Chowdhury and Tanmoy Chakraborty. 2019b. Cqasumm: Building references for community question answering summarization corpora. In Proceedings of the ACM India Joint International Conference on Data Science and Management of Data, pages 18–26. ACM.
  • Chowdhury et al. (2020) Tanya Chowdhury, Sachin Kumar, and Tanmoy Chakraborty. 2020. Neural abstractive summarization with structural attention. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI-20, pages 3716–3722. International Joint Conferences on Artificial Intelligence Organization. Main track.
  • DUC (2002) DUC. 2002. Document Understanding Conferences. [online] Available at: https://duc.nist.gov/. Accessed 1 Mar. 2020.
  • Erkan and Radev (2004) G. Erkan and D. R. Radev. 2004. Lexrank: Graph-based lexical centrality as salience in text summarization. Journal of Artificial Intelligence Research, 22:457–479.
  • Fabbri et al. (2019) Alexander R Fabbri, Irene Li, Tianwei She, Suyi Li, and Dragomir R Radev. 2019. Multi-news: a large-scale multi-document summarization dataset and abstractive hierarchical model. arXiv preprint arXiv:1906.01749.
  • Ganesan et al. (2010) Kavita Ganesan, ChengXiang Zhai, and Jiawei Han. 2010. Opinosis: a graph-based approach to abstractive summarization of highly redundant opinions. In Proceedings of the 23rd International Conference on Computational Linguistics, pages 340–348. Association for Computational Linguistics.
  • Gehrmann et al. (2018) Sebastian Gehrmann, Yuntian Deng, and Alexander Rush. 2018. Bottom-up abstractive summarization. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 4098–4109, Brussels, Belgium. Association for Computational Linguistics.
  • Gillick et al. (2008) Daniel Gillick, Benoit Favre, and Dilek Hakkani-Tür. 2008. The icsi summarization system at tac 2008. In Tac.
  • Grusky et al. (2018) Max Grusky, Mor Naaman, and Yoav Artzi. 2018. Newsroom: A dataset of 1.3 million summaries with diverse extractive strategies. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 708–719, New Orleans, Louisiana. Association for Computational Linguistics.
  • Hirao et al. (2004) Tsutomu Hirao, Takahiro Fukusima, Manabu Okumura, Chikashi Nobata, and Hidetsugu Nanba. 2004. Corpus and evaluation measures for multiple document summarization with multiple sources. In Proceedings of the 20th international conference on Computational Linguistics, page 535. Association for Computational Linguistics.
  • Hirao et al. (2017) Tsutomu Hirao, Masaaki Nishino, Jun Suzuki, and Masaaki Nagata. 2017. Enumeration of extractive oracle summaries. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers, pages 386–396, Valencia, Spain. Association for Computational Linguistics.
  • Jung et al. (2019) Taehee Jung, Dongyeop Kang, Lucas Mentch, and Eduard Hovy. 2019. Earlier isn’t always better: Sub-aspect analysis on corpus and system biases in summarization. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 3324–3335, Hong Kong, China. Association for Computational Linguistics.
  • Kryscinski et al. (2019) Wojciech Kryscinski, Nitish Shirish Keskar, Bryan McCann, Caiming Xiong, and Richard Socher. 2019. Neural text summarization: A critical evaluation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 540–551, Hong Kong, China. Association for Computational Linguistics.
  • Lebanoff et al. (2018) Logan Lebanoff, Kaiqiang Song, and Fei Liu. 2018. Adapting the neural encoder-decoder framework from single to multi-document summarization. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 4131–4141, Brussels, Belgium. Association for Computational Linguistics.
  • Lin (2004) Chin-Yew Lin. 2004. ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pages 74–81, Barcelona, Spain. Association for Computational Linguistics.
  • Mihalcea and Tarau (2004) Rada Mihalcea and Paul Tarau. 2004. TextRank: Bringing order into text. In Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, pages 404–411, Barcelona, Spain. Association for Computational Linguistics.
  • Nenkova and Passonneau (2004) Ani Nenkova and Rebecca Passonneau. 2004. Evaluating content selection in summarization: The pyramid method. In Proceedings of the Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics: HLT-NAACL 2004, pages 145–152, Boston, Massachusetts, USA. Association for Computational Linguistics.
  • Peyrard (2019a) Maxime Peyrard. 2019a. A simple theoretical model of importance for summarization. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 1059–1073.
  • Peyrard (2019b) Maxime Peyrard. 2019b. Studying summarization evaluation metrics in the appropriate scoring range. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 5093–5100.
  • See et al. (2017) Abigail See, Peter J. Liu, and Christopher D. Manning. 2017. Get to the point: Summarization with pointer-generator networks. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1073–1083, Vancouver, Canada. Association for Computational Linguistics.
  • TAC (2008) TAC. 2008. Text Analysis Conferences. [online] Available at: https://tac.nist.gov/. Accessed 1 Mar. 2020.
  • Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. CoRR, abs/1706.03762.