Enhancing Cross-Document Event Coreference Resolution by Discourse Structure and Semantic Information
Abstract
Existing cross-document event coreference resolution models,
which either compute mention similarity directly or enhance mention representation by extracting event arguments (such as location, time, agent, and patient),
lacking the ability to utilize document-level information. As a result,
they struggle to capture long-distance dependencies.
This shortcoming leads to their underwhelming performance in determining coreference for
the events where their argument information relies on long-distance dependencies.
In light of these limitations, we propose the construction of document-level Rhetorical Structure Theory (RST) trees and cross-document Lexical Chains to model the structural and semantic information of documents.
Subsequently, cross-document heterogeneous graphs are constructed and GAT is utilized to learn the representations of events.
Finally, a pair scorer calculates the similarity between each pair of events and co-referred events can be recognized using standard clustering algorithm.
Additionally, as the existing cross-document event coreference datasets are limited to English, we have developed a large-scale Chinese cross-document event coreference dataset to fill this gap, which comprises 53,066 event mentions and 4,476 clusters.
After applying our model on the English and Chinese datasets respectively,
it outperforms all baselines by large margins.
Keywords: Discourse Information, Cross-document Event Coreference, RST, Lexical Chains, English, Chinese
Enhancing Cross-Document Event Coreference Resolution by Discourse Structure and Semantic Information
Qiang Gao1, Bobo Li1, Zixiang Meng1, Yunlong Li1, Jun Zhou1, Fei Li1, Chong Teng1, Donghong Ji1 |
1Key Laboratory of Aerospace Information Security and Trusted Computing, Ministry of Education, |
School of Cyber Science and Engineering, Wuhan University |
[email protected], |
{boboli, zixiangmeng, yunlongli, j.zhou, lifei_csnlp, tengchong, dhji}@whu.edu.cn |
Abstract content
1. Introduction
Cross-document event coreference resolution (CDECR) is a critical task in natural language processing (NLP), exhibiting substantial applicability across various downstream tasks, including information extraction (Wen et al., 2021), document summarization (Li et al., 2020), question-answering (Khattab et al., 2021), etc. As shown in Figure 1(a), the primary goal of CDECR is to identify text references across multiple documents that relate to the same event. Given the necessity to comprehend an array of documents concurrently, CDECR is more challenging than within-document event coreference resolution (Dobrovolskii, 2021).


Recognizing these complexities, several methods have been proposed to address the challenges inherent to CDECR. Ravenscroft et al. (2021) utilized event mention representation to compute coreference scores, It often misjudges mentions that appear similar. Yu et al. (2022) proposed extracting event arguments to enhance the mention representation, but this approach is heavily reliant on the performance of the SRL (semantic role labeling) tool, unable to model discourse-level information. The most commonly used CDECR dataset ECB+ (Cybulska and Vossen, 2014), suffers from the limitations in terms of small data scale and event categorization, as it only addresses coreference within the events with the same type, which does not align well with real-world applications (Bugert et al., 2020). To address these limitations, Eirew et al. (2021) proposed a large-scale, domain-independent cross-document coreference resolution dataset called WEC-Eng. Following this work, we carried out our study on cross-document event coreference resolution. However, the baseline model presented in WEC-Eng only considers naive event mention features for coreference resolution, overlooking rich discourse-level information in the document. Therefore, we propose modeling discourse-level information to enhance cross-document event coreference resolution.
Discourse information provides a comprehensive view of events. On one hand, discourse structure can reveal the relationships between events and entities. The discourse structure often reflects certain logic clues such as causality and background, which help in understanding the overall narrative flow and the relative positions of events. As illustrated in Figure 1(b), the phrases like “four years later” and “before the head of Hong Kong Special Administrative Region Election" correspond to EDU④ and EDU⑤. EDU④ provides background information for EDU⑤. Moreover, for long documents, there are many long-distance dependencies between sentences. Due to the limitations of model input length, it is apt to cause information loss. Discourse structure information can shorten the distances between long-dependency sentences, providing a comprehensive understanding for the document.
Although discourse structure information is powerful, it cannot establish connections across documents. Therefore, semantic information beyond the discourse is required to establish fine-grained connections. Lexical information isolated in different documents sometimes cannot cover all the relevant information about the described events. Furthermore, the description of events in different documents may be quite different, resulting in the difficulty of event coreference resolution. For instance, in Figure 1(a), the same event might be represented differently such as “Hong Kong" and “Hong Kong Special Administrative Region", and the same meaning may be expressed using distinct words such as “first" and “inaugural". Consequently, there is a need to establish connections between lexical information across cross-document contexts, allowing for the direct mapping of event descriptions and event arguments.
In this work, we propose the DIE-EC (Discourse-Information-Enhanced Event Coreference) model, which leverages Rhetorical Structure Theory (RST) (Mann and Thompson, 1987) to construct document-level RST trees and cross-document lexical chains (Morris and Hirst, 1991), aiming to model both the structural and semantic information within and cross documents. Our model consists of (1) an encoder layer, which encodes the input document to obtain contextual representations; (2) a discourse information layer, which constructs document-level RST trees and cross-document-level lexical chains. The RST graph and lexical chains graph are processed using Graph Attention Networks (GAT) (Veličković et al., 2018); (3) a pair scorer, which processes the results from the GAT using an MLP (Multi-Layer Perceptron) for coreference determination, and finally clusters events using the agglomerative clustering algorithm. Besides the discourse information enhanced model, we also curate a Chinese cross-document event coreference dataset (WEC-Zh) due to its absence. The WEC-Zh consists of 53,066 event mentions, 4,476 event clusters and 8 event types. It is the first large-scale Chinese cross-document event coreference resolution dataset, promoting the research in the domains of Chinese and cross-lingual event coreference resolution. We experimented our model on both WEC-Eng and WEC-Zh datasets, and it outperformed all baselines by large margins in various evaluation metrics. To summarize, the contributions of this paper include:
-
1.
We propose the DIE-EC model, which involves the joint graph of RST and Lexical Chains, and utilizes GAT to extract discourse and cross-document features from it.
-
2.
We introduce a large-scale Chinese cross-document event coreference resolution dataset to facilitate related research.
-
3.
We conduct experiments on both WEC-Eng and our proposed WEC-Zh datasets. Our model achieves the state-of-the-art results on both datasets.111Our code and dataset are available at https://github.com/DIE-EC.
2. Related Work
Prior studies on cross-document event coreference resolution are performed around several datasets such as ECB+ (Cybulska and Vossen, 2014), Gun Violence Corpus (GVC) (Vossen et al., 2018) and WEC-Eng (Eirew et al., 2021). We will introduce these datasets as well as their related methods as below.
2.1. CDECR Datasets
The ECB+ dataset is a cross-document event coreference resolution dataset with a relatively small scale, comprising only 982 samples. It encompasses 26,712 coreference links between 6,833 event mentions and 69,050 coreference links between 8,289 entity mentions. Events are categorized by topic, comprising 43 topics in total. Each topic consists of different events that describe the same subject. Coreference annotations are specific to event categories and are only annotated within the same event category.
The GVC dataset pertains to the domain of gun violence. To overcome the costly endeavor of constructing large-scale datasets, the authors introduced an innovative semi-automatic method, known as structured-data-to-text (D2T). The D2T approach enables the creation of a vast amount of reference event data in a more efficient and highly consistent manner. GVC encompasses 5 gun-violence event classes (firing a gun, missing, hitting, injuring, death) and includes a total of 510 documents with 7,298 event mentions.
WEC-Eng is a large-scale cross-document coreference resolution dataset. To address the limitations of existing datasets, i.e., dataset scales are small or coreference relations are determined only within the same event category, the authors proposed a new way of constructing datasets. Instead of requiring extensive annotations, only the validation and test sets underwent manual verification. WEC-Eng comprises 43,672 documents with a total of 7,597 clusters. The coreference determination in this dataset is independent of the event category, making it more aligned with real-world applications. Although WEC-Eng has addressed the issues of existing datasets, it primarily focuses on English and lacks data in other languages. Hence, we have constructed a large-scale Chinese dataset.
2.2. CDECR Methods
Traditional coreference resolution methods mostly rely on mention representations for coreference score calculation. This approach often leads to a higher error rate when handling similar mentions. Therefore, Wu et al. (2020) proposed determining coreference in a question-answering format. it only determines coreference for event entities, leading to less-than-optimal accuracy. Held et al. (2021) introduced a coreference resolution model based on pre-filtering. This method reduces the computational cost by selecting only the top(n) mentions for coreference score calculation. However, the pre-filtering stage does introduce additional overhead. Similarly, Eirew et al. (2022) built an end-to-end model based on the Deep Passage Retrieval (DPR) model, which consists of a retriever and a reader model. Given a query paragraph, the retriever selects the top k most relevant candidate paragraphs from the entire paragraph corpus. Yu et al. (2022) proposed a method that enhances the semantic representation of mentions by extracting event argument information to improve the accuracy of coreference judgments. Miculicich and Henderson (2022) employed graph structures and multi-level iterative refinement to address coreference issues. The incorporation of syntactic parsing allows for fine-grained entity comparisons. However, the complexity of graph matrix construction is relatively high, making it unsuitable for long documents.
Existing methods primarily rely on calculating mention similarities to determine coreference without harnessing comprehensive document-level information. As a result, they tend to underperform in the scenarios where mentions exhibit certain similarities, or mentions have different descriptions but similar meanings. Therefore, we propose modeling document and cross-document information to enhance coreference resolution.
3. DIE-EC: A Discourse-Information -Enhanced Event Coreference Model for Cross-Document

DIE-EC leverages document-level RST trees and cross-document lexical chains to model the structural and semantic information of documents, as shown in Figure 2. For an input pair of documents, we first use Roberta (Liu et al., 2019) to extract the semantic representation of event mentions. Simultaneously, we build document-level RST trees for each of the documents and lexical chains for multiple documents. Then the RST trees are merged with lexical chains to obtain a comprehensive graph for the document pair. This combined graph is then fed into GAT for processing. Then we employ the node representations from GAT and mention representations to determine whether events are coreferential. Finally, the co-referenced events are clustered using the agglomerative clustering algorithm.
3.1. Encoder Layer
For input documents and , we separately employ Roberta to encode their contextual information, obtaining the contextual representations and for the documents. For event mentions and contained in the two documents, we first pad them to the same length and then use the concatenation of their token representations to represent them, denoted as and respectively.
3.2. Discourse Information Layer
3.2.1. Cross-document Graph Construction
RST Part
We employ RST to represent the structural information of a document. Based on rhetorical relation theory, we analyze the discourse structure and function of the document, partitioning it into EDUs to build a rhetorical structure tree. The nodes in the tree have two types: nucleus and satellite. The nucleus is a component containing the main information, while the satellite provides further explanation, supplementation, or modification to the nucleus. Different EDUs are connected by rhetorical relations. RST plays a crucial role in discourse comprehension. By analyzing the text’s structure and rhetorical techniques, aids the model in better understanding the article’s intent, the organization of information, and the relationship between various parts. We adopt the method proposed by Zhang et al. (2020) to construct RST for the input documents, for Chinese data, we use the method proposed by Kong and Zhou (2017) based on CDTB (Chinese Discourse Treebank) for construction. Let the nodes where event mentions and are located be represented as and , respectively. In the RST tree, node relationships go from the satellite pointing to the nucleus or from the nucleus pointing to another nucleus. When converting the RST tree into a graph, we determine the direction of the edges in the graph based on the relationships between the nodes in the tree. Since the positions of two nucleus nodes are equivalent, we set the edge between the two nucleus nodes to be bidirectional.
Lexical Chain Part
We utilize lexical chains to furnish semantic information of the document. Lexical chains offer insights that can assist systems in grasping the semantic connections between lexemes, thereby extracting information more precisely. It permits the system to account for semantic similarity, ensuring a more accurate determination of the relationships between entities or events mentioned within the document. Following the approach proposed by Ruas et al. (2020), we construct lexical chains for repetition, synonyms, words close semantic relationships and temporal relationships. For the two input documents, we build cross-document lexical chains , where denotes the -th lexical chain constructed and symbolizes the word within the chain. Let the lexical chains that contain the words from the event mentions of the two documents be and , with starting lexemes and respectively. We construct a lexical chains graph with each word serving as a node, and all edges in the graph being bidirectional.
Merging RST and Lexical Chain
As only the leaf nodes in the RST graph contain the semantic information of the input document, we connect these leaf nodes with the nodes in the lexical chains graph to effectuate their fusion. The starting node of the lexical chain graph is connected to the EDU containing the starting lexeme; that is, is linked to and is linked to . If different EDUs contain lexical chains, an edge is established between the corresponding EDUs. After merging, we obtain a connected graph, simultaneously endowed with both the document’s structural and semantic information.
3.2.2. Graph Representation Learning
For the merged graph , we employ the GAT model to learn about the semantics and rhetorical relationships in the text, aiming to extract structural and semantic information. The GAT updates the feature representation of a target node by aggregating the features of its neighboring nodes. This approach eliminates the need to access information from the entire graph, enhancing the model’s generalization capabilities. Moreover, through the attention mechanism, GAT can learn different attention weights for each neighboring node. This facilitates an effective aggregation of the neighboring nodes’ features, leading to enhanced feature extraction capability for the target node.
For the graph consisting of nodes We use the representations derived from the encoder to initialize the nodes as . After processing through GAT, we obtain representations for the EDU nodes as , where and represents the dimension of node representations. can be calculated using the following formula:
(1) |
where is the number of attention heads, represents the Relu activation function, represents the set of neighboring nodes of node , is weight matrix for attention head. represents the attention weight between node and node , which can be calculated as follows:
(2) |
where is the attention parameter vector, represents the concatenation operation of vectors. We use and to represent the EDU nodes where event mentions are located.
3.3. Pair Scorer
Computing Coreference Score
After obtaining the representations, we proceed to calculate the coreference score. We fuse the event mention representations and with the GAT node representations and to derive the representation for the event mention pair, denoted as:
(3) |
Once we obtain the representation of the event mention pair, we employ a 3-layer MLP wherein the last layer performs a binary classification. Through the sigmoid function, we derive the coreference score for the mention pair as:
(4) |
Clustering
All pairs with a probability lower than 0.5 are eliminated. Following Barhom et al. (2019), we use an agglomerative clustering algorithm for clustering. It treats each data point as an initial cluster and progressively merges them into larger clusters. During the merging process, we use the group average, which measures the similarity between clusters based on the average distance of all elements within a cluster. The clustering process concludes when the similarity between all clusters falls below a predefined threshold of 0.7.
4. WEC-Zh: Wikipedia Event Coreference Dataset for Chinese
In this paper, we also curate a cross-document event coreference resolution dataset, WEC-Zh, to compensate the shortage of Chinese resource in this domain. We will showcase the construction process and dataset statistics of the WEC-Zh dataset.
4.1. Construction Process
Our dataset originates from the Wikipedia Chinese data. Following the precedent set by WEC-Eng (Eirew et al., 2021), our data was collected by aggregating anchor texts of Wikipedia links that point to the same Wikipedia concept. In other words, we consider links in an article that point to another article about the same real-world subject as being co-referential. Upon obtaining the raw data, we executed a series of filtering operations: (1) Removing the mentions where event links were placed on parameter information (such as time, location, person and other entities that cannot represent events). (2) Deleting non-Chinese event mentions.
After the filtering operations, we converted Traditional Chinese to Simplified Chinese, ensuring that the resulting dataset consisting exclusively of Simplified Chinese data. To guarantee the quality of the dev and test sets, we conducted a two-person cross-validation on the dev and test datasets, eliminating noisy data and mentions whose coreferences could not be determined based on context. Figure 3 illustrates an example of WEC-zh.
To further assess the quality of annotators’ labeling, we employed Fleiss’ Kappa algorithm (Fleiss, 1971) to calculate annotation consistency. The Kappa consistency value is 0.71, indicating that our annotations achieves a acceptable agreement rate and showing the descent quality of our dataset.
4.2. Dataset Statistics
As shown in Table 1, our dataset encompasses over 50,000 event mentions, which includes a certain number of singleton samples. Unlike traditional datasets like ECB+ (Cybulska and Vossen, 2014) that are clustered by event categories and then coreference resolution is performed within each category, our dataset determines event coreference without such constraint. To ensure domain diversity of the dataset, we tallied the numbers of documents with different event types, as presented in Table 2. For more detailed statistical information, please refer to Appendix A.
Mentions | Clusters | Single Clusters | Ambiguity | Diversity | |
Train | 49,861 | 3,855 | 1,110 | 3.31 | 2.35 |
Dev | 1,538 | 297 | 84 | 1.82 | 1.69 |
Test | 1,667 | 324 | 76 | 1.83 | 1.56 |
Train | Dev | Test | |
ATTACK_EVENT | 31,569 | 886 | 917 |
SPORT_EVENT | 4,015 | 150 | 259 |
EVENT_UNK | 6,541 | 144 | 122 |
ELECTION_EVENT | 4,605 | 118 | 161 |
GENERAL_EVENT | 1,426 | 155 | 131 |
DISASTER_EVENT | 911 | 32 | 40 |
ACCIDENT_EVENT | 439 | 39 | 20 |
AWARD_EVENT | 197 | 12 | 15 |
OTHERS | 158 | 2 | 2 |

5. Experiments
MUC | CoNLL | |||||||||||||
R | P | R | P | R | P | |||||||||
Lemma | 85.5 | 79.9 | 82.6 | 74.5 | 32.8 | 45.5 | 25.9 | 39.4 | 31.2 | 53.1 | ||||
WEC-Eng | 78 | 83.6 | 80.7 | 66.1 | 55.3 | 60.2 | 53.4 | 40.3 | 45.9 | 62.3 | ||||
Our model | 78.2 | 85.8 | 81.8 | 69.6 | 62.4 | 65.8 | 58.9 | 39.5 | 47.3 | 65.0 |
MUC | CoNLL | |||||||||||||
R | P | R | P | R | P | |||||||||
Lemma | 80.5 | 45.0 | 57.7 | 84.3 | 47.1 | 60.4 | 37.3 | 52.4 | 43.6 | 53.9 | ||||
WEC-Eng | 91.1 | 54.2 | 68.0 | 91.9 | 52.3 | 66.7 | 39.0 | 76.3 | 51.6 | 62.1 | ||||
Our model | 90.3 | 57.9 | 70.6 | 90.8 | 60.2 | 72.4 | 46.7 | 76.0 | 57.8 | 66.9 |
5.1. Experiment Setting
Baselines We conducted experiments on both WEC-Eng and WEC-Zh. The lemma model and WEC-Eng model proposed by Eirew et al. (2021) are used as baselines for both datasets. In the lemma baseline, if two event mentions share the same syntactic-head lemma, they are considered co-referential. To ensure a balanced sample, we maintained a positive-to-negative sample ratio of 1:10 for the training set. There was no such control for dev and test sets. Positive samples consist of all mention pairs that belong to the same cluster.
Evaluation Metrics We utilized the agglomerative clustering algorithm for clustering and reported P, R and scores on the MUC, , CEAF, and CoNLL metrics for the clustering results.
Training Details Our experiments were conducted on two RTX 3090 GPUs. The initial learning rate was set to 1e-5 with a warm-up phase. The batch size was set to 128. We utilized the Adam optimizer as the optimization algorithm and set a weight decay parameter of 0.01 to prevent overfitting. The entire process was run for 10 iterations.
5.2. Main Results
WEC-Eng The experimental results of our model on WEC-Eng are shown in Table 3. The results for the Lemma baseline and WEC-Eng baseline are sourced from the WEC-Eng. Our model outperforms the WEC-Eng baseline by 5.6 points of , by 2.7 points of CoNLL, and also excels in the MUC metric compared to the WEC-Eng baseline. The WEC-Eng baseline has a higher error rate for mentions that are highly similar but not coreferential, whereas our model achieves a higher accuracy rate. This evidences that our model’s adoption of RST and lexical chains for modeling the structural and semantic information of documents genuinely aids in better determining the coreference information of event mentions. The erroneous case studies are provided in Appendix B.
WEC-Zh The experimental results on WEC-Zh are shown in Table 4. Our model outperforms the WEC-Eng model in terms of score across all metrics, with score of surpassing the WEC-Eng baseline by 5.7, and score of CEAF exceeding the WEC-Eng baseline by 6.2. This suggests that, for the WEC-Zh dataset, our approach of constructing document-level RST trees and cross-document lexical chains better extracts event-related information from discourse-level text, to some extent reducing potential ambiguities and enhancing model performance. Compared to the English dataset, our model achieved a more significant improvement on the Chinese dataset. This could be attributed to the fact that, in contrast to English, Chinese exhibits a wide variety of expressions for the same events, along with abbreviations and pronouns. This necessitates the model to establish correlations at the discourse-level for extracting useful information.
MUC | CoNLL | |||||||||||||
R | P | R | P | R | P | |||||||||
Our-model | 78.2 | 85.8 | 81.8 | 69.6 | 62.4 | 65.8 | 58.9 | 39.5 | 47.3 | 65.0 | ||||
- RST | 78.1 (-0.1) | 84.4 (-1.4) | 81.1 (-0.7) | 68.4 (-1.2) | 60.1 (-2.3) | 64.0 (-1.8) | 57.1 (-1.8) | 39.4 (-0.1) | 46.6 (-0.7) | 63.9(-1.1) | ||||
- Lexical chains | 77.9(-0.3) | 83.9 (-1.9) | 80.8 (-1.0) | 64.2 (-5.2) | 57.9 (-4.5) | 60.9 (-4.9) | 53.6 (-5.3) | 39.2 (-0.3) | 45.3 (-2.0) | 62.4 (-2.6) |
MUC | CoNLL | |||||||||||||
R | P | R | P | R | P | |||||||||
Our-model | 90.3 | 57.9 | 70.6 | 90.8 | 60.2 | 72.4 | 46.7 | 76.0 | 57.8 | 66.9 | ||||
- RST | 90.2 (-0.1) | 55.3 (-2.6) | 68.6 (-2.0) | 89.8 (-1.0) | 55.8 (-4.4) | 68.8 (-3.6) | 42.6 (-0.1) | 75.8 (-0.2) | 54.5 (-3.3) | 64.0(-2.9) | ||||
- Lexical chains | 90.1(-0.2) | 54.7 (-3.2) | 68.1 (-2.5) | 87.4 (-3.4) | 54.1 (-6.1) | 66.8 (-5.6) | 40.5 (-6.2) | 75.9 (-0.1) | 53.0 (-5.0) | 62.6 (-4.3) |
5.3. Ablation Studies
To further investigate the contribution of RST trees and lexical chains for cross-document coreference resolution, we conducted a series of ablation experiments, and the experimental results are shown in Table 5 and 6.
From the ablation study results, it is evident that removing either RST trees or lexical chains leads to a decrease in model performance. This indicates that discourse information indeed enhances the model’s coreference resolution capabilities. This is attributed to the fact that the discourse structure provides the logical relationships between an event and its surrounding context, while the discourse semantics captures the continuity of the event and the evolutionary process of related information. These two types of information complement each other, aiding the system in a deeper understanding and comparison of event descriptions across different documents.
Additionally, a significant performance drop is observed upon removing the lexical chains, suggesting that even though the RST trees provide structural information of the discourse, the contribution of lexical chains to the model’s performance is significant. This might indicate that in cross-document event coreference resolution tasks, deep semantic coherence and topic continuity play a pivotal role. While the structural relationships provided by RST trees do assist in coreference resolution, they are not as direct and crucial as the semantic coherence captured by lexical chains. When dealing with events across documents, recognizing and matching the same or similar semantic content might be more important than just understanding the structure and narrative flow.
5.4. In-depth Analysis
To further analyze the effects of our RST trees and lexical chains on different samples, we conducted two experiments: (1) The impact of lexical chains with different lexical overlap rates. The lexical overlap rate is calculated as the ratio of overlapping words between two documents to the average total number of words in both documents. (2) The impact of RST tree with different document lengths. The length of a document is defined as the number of words it contains.
(1) Different Lexical Overlap Rates
We hypothesized that events which are likely to be coreferent often contain a higher degree of lexical overlap between documents, as their descriptions share similar content. In such cases, the model can make judgments more easily, and the role of lexical chains is relatively minor. Conversely, for non-coreferent events, where descriptions differ or are only partially similar, the lexical overlap between documents is lower, leading to potential confusion for the model. In these cases, lexical chains play a more critical role in connecting vocabulary across documents to assist the model in making judgments. To validate our hypothesis, we computed the lexical overlap rates for all sample pairs and conducted ablation experiments, measuring scores of . The experimental results are shown in Figure 4. It is evident that when the lexical overlap rate exceeds 50%, it is easier to determine co-reference, removing lexical chains does not lead to a significant decrease in the model’s performance. However, when the overlap rate falls below 30%, the model’s performance starts to decline substantially. Specifically, when it drops below 10%, scores decrease by 8.5 on WEC-Eng, and by 10.1 on WEC-Zh. This validates our hypothesis that for texts with a low lexical overlap rate, lexical chains play a significant role in co-reference resolution.
MUC | CoNLL | |||||||||||||
R | P | R | P | R | P | |||||||||
PairwiseRL | 88.1 | 85.1 | 86.6 | 86.1 | 84.7 | 85.4 | 83.1 | 79.6 | 81.3 | 84.4 | ||||
DRS CD-ECR | 88.6 | 85.9 | 87.2 | 87.8 | 85.4 | 86.6 | 82.8 | 83.7 | 83.2 | 85.7 | ||||
Our model | 88.1 | 87.6 | 87.8 | 86.9 | 85.8 | 86.3 | 81.7 | 83.2 | 82.4 | 85.5 |
MUC | CoNLL | |||||||||||||
R | P | R | P | R | P | |||||||||
Our model | 88.1 | 87.6 | 87.8 | 86.9 | 85.8 | 86.3 | 81.7 | 83.2 | 82.4 | 85.5 | ||||
- RST | 86.3 (-1.8) | 84.2 (-3.4) | 85.2 (-2.6) | 85.1 (-1.8) | 83.5 (-2.3) | 84.3 (-2.0) | 79.8 (-1.9) | 82.1 (-1.1) | 81.0 (-1.4) | 83.5(-2.0) | ||||
- Lexical chains | 84.9(-3.2) | 83.7 (-3.9) | 84.3 (-3.5) | 84.2 (-2.7) | 82 (-3.8) | 83.1 (-3.2) | 78.3 (-3.4) | 80.4 (-2.8) | 79.3 (-3.1) | 82.2 (-3.3) |
(2) Different Document Lengths
Longer documents often contain a wealth of contextual dependencies. Previous methods, due to direct context truncation, may overlook this information, resulting in some loss of information. We hypothesize that RST trees can model the discourse structure of long documents, extracting long-range dependencies that enhance coreference resolution. To validate our hypothesis, we conducted experiments on document pairs of varying lengths. The experimental results are shown in Figure 5. It is evident that when the document length is less than 512, removing the RST model only results in a slight performance drop. However, when the document length exceeds 1024, the F1 scores decrease by 6.3 on WEC-Eng, by 7.8 on WEC-Zh. This demonstrates that RST is capable of modeling long documents and capturing long-range dependencies, thus validating our hypothesis.
5.5. ECB+ Experiment
Besides the experiments on WEC-Eng and WEC-Zh, we also conducted experiments on ECB+ (Cybulska and Vossen, 2014) because it is widely used for evaluating models of cross-document event coreference resolution.
5.5.1. Experimental Settings
When forming mention pairs from the ECB+ dataset, all positive samples are paired with each other, and all possible negative samples within the same subtopic are also paired with each other. The final distribution of these pairs is as follows:
Dev set: Positives=5,881, Negatives=50,653
Test set: Positives=6,889, Negatives=87,053
Train set: Positives=14,944, Negatives=170,549
5.5.2. Experimental Results
The results of the baseline comparison experiment and ablation studies are shown in Table 7 and 8 respectively. From the experimental results, it can be seen that our model achieves comparable performance with the SOTA model DRS CD-ECR, which is reasonable because both of our and their models utilize discourse structural knowledge. The performance differences may be due to the implementation details such as using different RST parsers or the methods of leveraging RST. Compared to PairwiseRL, our model surpasses it in MUC by 1.2, B3 by 0.9, CEAF by 1.1 and CoNLL by 1.1. This demonstrates that our approach of modeling the global structural and semantic information of documents is more effective than enhancing event mention representations by extracting event parameters.
6. Conclusion
In this study, we introduce an improved method for cross-document event coreference resolution by incorporating discourse-level information. By constructing document-level RST trees and cross-document Lexical chains, we can more effectively capture the structural and semantic information within documents, especially long-distance dependencies. Moreover, we have constructed a large-scale cross-document event coreference dataset for the Chinese context, filling a gap in current research. This work offers valuable resources and methodologies for further studies on cross-document event coreference resolution.
7. Bibliographical References
- Barhom et al. (2019) Shany Barhom, Vered Shwartz, Alon Eirew, Michael Bugert, Nils Reimers, and Ido Dagan. 2019. Revisiting joint modeling of cross-document entity and event coreference resolution. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 4179–4189, Florence, Italy. Association for Computational Linguistics.
- Bugert et al. (2020) Michael Bugert, Nils Reimers, Shany Barhom, Ido Dagan, and Iryna Gurevych. 2020. Breaking the subtopic barrier in cross-document event coreference resolution. In Proceedings of Text2Story - Third Workshop on Narrative Extraction From Texts co-located with 42nd European Conference on Information Retrieval, Text2Story@ECIR 2020, Lisbon, Portugal, April 14th, 2020 [online only], volume 2593 of CEUR Workshop Proceedings, pages 23–29. CEUR-WS.org.
- Chen et al. (2023) Xinyu Chen, Sheng Xu, Peifeng Li, and Qiaoming Zhu. 2023. Cross-document event coreference resolution on discourse structure. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 4833–4843, Singapore. Association for Computational Linguistics.
- Cybulska and Vossen (2014) Agata Cybulska and Piek Vossen. 2014. Using a sledgehammer to crack a nut? lexical diversity and event coreference resolution. In Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC’14), pages 4545–4552, Reykjavik, Iceland. European Language Resources Association (ELRA).
- Dobrovolskii (2021) Vladimir Dobrovolskii. 2021. Word-level coreference resolution. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 7670–7675, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Eirew et al. (2022) Alon Eirew, Avi Caciularu, and Ido Dagan. 2022. Cross-document event coreference search: Task, dataset and modeling. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 900–913, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- Eirew et al. (2021) Alon Eirew, Arie Cattan, and Ido Dagan. 2021. WEC: Deriving a large-scale cross-document event coreference dataset from Wikipedia. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2498–2510, Online. Association for Computational Linguistics.
- Fleiss (1971) JL Fleiss. 1971. Measuring nominal scale agreement among many raters. Psychological bulletin, 76(5):378—382.
- Held et al. (2021) William Held, Dan Iter, and Dan Jurafsky. 2021. Focus on what matters: Applying discourse coherence theory to cross document coreference. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 1406–1417, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Khattab et al. (2021) Omar Khattab, Christopher Potts, and Matei Zaharia. 2021. Baleen: Robust multi-hop reasoning at scale via condensed retrieval. Advances in Neural Information Processing Systems, 34:27670–27682.
- Kong and Zhou (2017) Fang Kong and Guodong Zhou. 2017. A cdt-styled end-to-end chinese discourse parser. ACM Trans. Asian Low-Resour. Lang. Inf. Process., 16(4).
- Li et al. (2020) Wei Li, Xinyan Xiao, Jiachen Liu, Hua Wu, Haifeng Wang, and Junping Du. 2020. Leveraging graph to improve abstractive multi-document summarization. arXiv preprint arXiv:2005.10043.
- Liu et al. (2019) Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach.
- Mann and Thompson (1987) William Mann and Sandra Thompson. 1987. Rhetorical Structure Theory: A Theory of Text Organization.
- Miculicich and Henderson (2022) Lesly Miculicich and James Henderson. 2022. Graph refinement for coreference resolution. In Findings of the Association for Computational Linguistics: ACL 2022, pages 2732–2742, Dublin, Ireland. Association for Computational Linguistics.
- Morris and Hirst (1991) Jane Morris and Graeme Hirst. 1991. Lexical cohesion computed by thesaural relations as an indicator of the structure of text. Computational Linguistics, 17(1):21–48.
- Ravenscroft et al. (2021) James Ravenscroft, Amanda Clare, Arie Cattan, Ido Dagan, and Maria Liakata. 2021. CD^2CR: Co-reference resolution across documents and domains. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pages 270–280, Online. Association for Computational Linguistics.
- Ruas et al. (2020) Terry Ruas, Charles Henrique Porto Ferreira, William I. Grosky, Fabrício Olivetti de França, and Debora Maria Rossi de Medeiros. 2020. Enhanced word embeddings using multi-semantic representation through lexical chains. ArXiv, abs/2101.09023.
- Veličković et al. (2018) Petar Veličković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua Bengio. 2018. Graph attention networks.
- Vossen et al. (2018) Piek Vossen, Filip Ilievski, Marten Postma, and Roxane Segers. 2018. Don’t annotate, but validate: a data-to-text method for capturing event data. In Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018), Miyazaki, Japan. European Language Resources Association (ELRA).
- Wen et al. (2021) Haoyang Wen, Ying Lin, Tuan Lai, Xiaoman Pan, Sha Li, Xudong Lin, Ben Zhou, Manling Li, Haoyu Wang, Hongming Zhang, Xiaodong Yu, Alexander Dong, Zhenhailong Wang, Yi Fung, Piyush Mishra, Qing Lyu, Dídac Surís, Brian Chen, Susan Windisch Brown, Martha Palmer, Chris Callison-Burch, Carl Vondrick, Jiawei Han, Dan Roth, Shih-Fu Chang, and Heng Ji. 2021. RESIN: A dockerized schema-guided cross-document cross-lingual cross-media information extraction and event tracking system. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Demonstrations, pages 133–143, Online. Association for Computational Linguistics.
- Wu et al. (2020) Wei Wu, Fei Wang, Arianna Yuan, Fei Wu, and Jiwei Li. 2020. CorefQA: Coreference resolution as query-based span prediction. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 6953–6963, Online. Association for Computational Linguistics.
- Yu et al. (2022) Xiaodong Yu, Wenpeng Yin, and Dan Roth. 2022. Pairwise representation learning for event coreference. In Proceedings of the 11th Joint Conference on Lexical and Computational Semantics, Seattle, Washington. Association for Computational Linguistics.
- Zhang et al. (2020) Longyin Zhang, Yuqing Xing, Fang Kong, Peifeng Li, and Guodong Zhou. 2020. A top-down neural architecture towards text-level parsing of discourse rhetorical structure. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 6386–6395, Online. Association for Computational Linguistics.
8. Appendices
Appendix A Appendix: The details of WEC-Zh dataset
A.1. Event Mention Details
To investigate the distribution of coreference clusters, we counted the number of clusters corresponding to different mention scopes, as depicted in Table 9.
We also conducted headword statistics for event mentions and performed Named Entity Recognition (NER) tagging on the headwords, as shown in Table 10. Additionally, we carried out Part-of-Speech (POS) tagging on the headwords, and the results are presented in Table 11.
To illustrate the diversity of events in the dataset, we conducted a statistical analysis of the unique mention count and unique head count across three partitions, as presented in Table 12.
Train | Dev | Test | |
>=2 | 2,745 | 213 | 248 |
>=5 | 1,477 | 138 | 144 |
>=10 | 793 | 64 | 66 |
>=20 | 459 | 0 | 0 |
>=50 | 163 | 0 | 0 |
Train | Dev | Test | |
UNK | 23,749 | 1,021 | 1,067 |
EVENT | 22,951 | 431 | 481 |
DATE | 787 | 7 | 28 |
FAC | 714 | 18 | 5 |
WORK_OF_ART | 491 | 31 | 23 |
LOC | 432 | 9 | 24 |
GPE | 431 | 3 | 9 |
ORG | 129 | 15 | 14 |
CARDINAL | 53 | 1 | 5 |
ORDINAL | 48 | 0 | 0 |
PERSON | 40 | 1 | 6 |
LAW | 21 | 0 | 4 |
QUANTITY | 10 | 1 | 1 |
NORP | 4 | 0 | 0 |
TIME | 1 | 0 | 0 |
Train | Dev | Test | |
NOUN | 42,522 | 1,322 | 1,412 |
VERB | 4,470 | 191 | 209 |
PROPN | 2,337 | 15 | 24 |
PUNCT | 173 | 6 | 14 |
ADV | 19 | 0 | 7 |
NUM | 315 | 4 | 1 |
DET | 13 | 0 | 0 |
PRON | 4 | 0 | 0 |
OTHERS | 3 | 0 | 0 |
ADJ | 2 | 0 | 0 |
PART | 2 | 0 | 0 |
ADP | 1 | 0 | 0 |
Train | Dev | Test | |
unique-mentions | 9,402 | 563 | 616 |
unique-heads | 2,119 | 230 | 239 |
WEC-Eng | WEC-Zh | |
Our-model | 65.0 | 66.9 |
Background | 64.7 | 66.8 |
Elaboration | 64.8 | 65.5 |
Explanation | 64.6 | 65.6 |
Enablement | 64.8 | 66.7 |
Cause | 64.7 | 66.9 |
Contrast | 64.9 | 66.1 |
Comparison | 64.7 | 66.4 |
Evaluation | 64.6 | 66.8 |
Summary | 64.9 | 66.2 |
Temporal | 64.8 | 66.7 |
A.2. Sample Demonstration
We have provided some Chinese dataset samples in Table 14.
Event mention and context | G | W | O | -R | -L |
公元290年,晋武帝去世,晋惠帝司马衷继位,司马衷“不慧”,至皇后贾南风专政,后八王之乱爆发,晋惠帝被毒杀, 晋怀帝司马炽继位,八王之乱致西晋国力大损,各游牧民 族纷纷起兵入侵,并建立政权,引发五胡乱华,公元311年 汉赵皇帝刘聪攻入洛阳,晋怀帝被掳后被杀,晋愍帝司马邺在长安继位,不久长安亦被汉赵攻陷,西晋亡。 In 290 AD, Emperor Wu of Jin passed away, and Emperor Hui of Jin, Sima Zhong, succeeded to the throne. Sima Zhong was “not wise” and ruled by Empress Jia Nanfeng. Later, the Eight Kings Rebellion broke out, and Emperor Huiof Jin was poisoned, Emperor Huai of Jin, Sima Chi, succeeded to the throne, and the Rebellion of the Eight Kings caused a great loss of national strength in the Western Jin Dynasty. Various nomadic ethnic groups rose up to invade and establish political power, triggering Upheaval of the Five Barbarians. In 311 AD Emperor Liu Cong of Han and Zhao invaded Luoyang, Emperor Huai of Jin was captured and killed, and Emperor Min of Jin, Sima Ye, succeeded to the throne in Chang an. Shortly after, Chang an was also captured by Han and Zhao, and the Western Jin Dynasty perished. 西晋末年,中原发生八王之乱,接着“匈奴、鲜卑、羯、羌、氐五个部落建立非汉族政权”,史称“五胡乱华”,为避 战乱,大量人口举族渡江南迁,东晋在江南各地设侨州、侨郡,安置流民。京口作为重要渡口,此侨置南徐州、南兖州 以及南东海、南琅琊、南兰陵、南濮阳等18郡,县治多至60多个,移民数量远超土著。 In the late years of the Western Jin Dynasty, the Eight Kings Rebellion occurred in the Central Plains. Subsequently, five tribes of the Hu people, including the Xiongnu, Xianbei, Jie, Qiang, and Di tribes, established non-Han political power, known as the “Upheaval of the Five Barbarians” in history. To avoid war, a large number of people crossed the Yangtze River and migrated south. The Eastern Jin Dynasty established overseas Chinese prefectures and counties in various parts of the Jiangnan region to resettle refugees. As an important ferry, Jingkou is home to 18 counties including Nanyanzhou, and Nandonghai, Nanlangya, Nanlanling, and Nanpuyang. There are as many as 60 county governments, and the number of immigrants far exceeds that of the indigenous people. | ✓ | ✗ | ✓ | ✓ | ✗ |
1996年7月,葛菲/顾俊的组合代表中国参加在美国亚特兰大举行的奥运会羽毛球比赛女子双打项目;她们在晋级过 程中所向披靡,四场赛事仅仅失了39分。决赛中,葛/顾面对当时世界排名第一的韩国组合吉永雅/张惠玉,结果葛/顾 只花了36分钟,便以2比0(15-5、15-5)轻松击败对手,夺得中国首面的羽毛球奥运金牌。 In July 1996, the combination of Ge Fei and Gu Jun represented China in the women’s doubles event of the Olympic badminton competition held in Atlanta, USA; They were unbeatable in the promotion process, losing only 39 points in four matches. In the final, Ge/Gu faced the then-world number one South Korean Ji Yongya/Zhang Huiyu, but in just 36 minutes, Ge/Gu easily defeated their opponents 2-0 (15-5, 15-5) to win China’s first badminton Olympic gold medal. 奥运会后,葛菲/顾俊接连赢得1997年世界羽毛球锦标赛冠军、1998年亚洲运动会羽毛球比赛金牌和1999年世界羽 毛球锦标赛冠军等等;更在四年半间以无敌姿态取得全胜佳绩,连胜纪录长达100场左右。 After the Olympics, Ge Fei/Gu Jun successively won the championship of the 1997 World Badminton Championships, the gold medal of the 1998 Asian Games Badminton Championships, and the championship of the 1999 World Badminton Championships; In four and a half years, he achieved a total victory with an invincible posture, with a winning streak of about 100 games. | ✗ | ✗ | ✓ | ✗ | ✗ |
辽宁红沿河核电厂位于大连市瓦房店的温坨子村,厂区三面环海,一面与陆地接壤。复州河在厂址以南约20公里处入海。 厂址区域地处华北地震区,绝大部分位于郯庐断裂带的北半部,地震活动,历史上最大地震为1975年海城7.3级地震, 震中距厂址约143公里。 The Liaoning Hongyanhe Nuclear Power Plant is located in Wentuozi Village, Wafangdian, Dalian City. The factory area is surrounded by the sea on three sides and borders land on one side. The Fuzhou River flows into the sea about 20 kilometers south of the factory site. The factory site area is located in the North China earthquake zone, with the majority located in the northern half of the Tanlu fault zone. It is seismically active, with the largest earthquake in history being the 1975 Haicheng 7.3 magnitude earthquake, with an epicenter approximately 143 kilometers from the factory site. 1975年2月4日凌晨,辽宁人民广播电台直接播出地震预报。”辽宁省省委做出指示:“从当天晚上起,辽南地区海城、 营口两县,所有人员都不要住在室内,生产队的大牲口、农业机械都要拉到室外。各级干部、党员、民兵全部下去,挨家 挨户动员老百姓。在生产队和城镇的居民区,用大喇叭广播动员 群众。”当日19点36分,海城地区爆发7.3级强烈地震, 由于发布消息及时,海城伤亡人数为:伤一万余人,死一千三百余人,损失较小,挽救了10多万人的生命。 On the early morning of February 4, 1975, Liaoning People’s Broadcasting Station directly broadcasted earthquake predictions The Liaoning Provincial Party Committee has issued instructions: "Starting from the evening of that day, all personnel in Haicheng and Yingkou counties in the southern region of Liaoning should not live indoors, and the large livestock and agricultural machinery of the production team should be pulled outdoors. Cadres, party members, and militia at all levels should go down and mobilize the people door-to-door. In the residential areas of the production team and the town, loudspeakers should be used to mobilize the people." At 19:36 on the same day, a strong earthquake measuring 7.3 occurred in Haicheng area, Due to the timely release of information, the number of casualties in Haicheng was: over 10,000 injured and over 1,300 dead, with relatively small losses and saving over 100,000 lives. | ✓ | ✗ | ✓ | ✓ | ✗ |
Appendix B Appendix: Error analysis
B.1. Error Case Studies
To further explore the performance of our method on different specific examples and to intuitively demonstrate the assistance of RST and lexical chains in analyzing challenging event coreference cases, we selected several categories of examples from the error cases of the WEC-Eng model for experimentation with our model. The selected examples include: ① Highly similar mentions, ② Mentions with different semantic expressions but the same meaning, ③ Mentions with remote dependencies on event parameters. The experimental results are shown in Table 15.
The results reveal that for highly similar mentions like Eurovision Song Contest 1980 and Eurovision Song Contest 1969, lexical chains determine the correctness of coreference judgments. This is because the lexical chains, which include time-related information such as (1980-1969) and location information like (The Hague-Madrid), provide direct evidence for non-coreference of the events. Additionally, the structural information related to “recorded by" and “written by" also provides some corroborative information but does not play a decisive role.
For mentions with different expressions but the same semantics, such as Assassination of Robert F. Kennedy and shooting, there is no apparent rhetorical logical relationship between these events. However, there are abundant time, location, and proper noun expressions, so lexical chains play a crucial role in coreference judgment for these events. The model can construct the following lexical chains: (Senator Robert F.Kennedy – Kennedy – Robert F.Kennedy), (June 5, 1968 – June 5, 1968), (the kitchen of the Ambassador Hotel in Los Angeles – The Ambassador Hotel in Los Angeles), (1968 Democratic presidential primary in California – Democratic 1968 California U.S. presidential primary). Although the event descriptions are different, there is a significant amount of repeated content related to event mentions, making it easier for the model to determine coreference for these events.
For events with remote dependencies in mentions, such as 42nd American Music Awards ceremony and American Music Awards, there is a clear temporal evolution of the event parameters, for example, (November 24, 2014 One day before) – (in November 16, 2014 On November 23 On November 28). Therefore, RST can accurately capture the temporal evolution relationship between events. Combined with lexical chains information like (song – music – band), it enables the model to establish a connection between the two events, providing support for the model’s judgment.
Event mention and context | G | W | O | -R | -L |
Kennedy (known as Jean Kennedy Smith following her 1956 marriage to Stephen Edward Smith) was intricately involved with the political career of her older brother John. She worked on his 1946 Congressional campaign, his 1952 Senate campaign, and ultimately his presidential campaign in 1960. She and her siblings helped Kennedy knock on doors in primary states like Texas and Wisconsin and on the campaign trail played the role of sister more than volunteer, citing her parent’s family lesson of working together for something. Smith and her husband was present at The Ambassador Hotel in Los Angeles on June 5, 1968, during the Assassination of Robert F. Kennedy after he had won the Democratic 1968 California U.S. presidential primary. Bobby is a 2006 American drama film written and directed by Emilio Estevez, and starring an ensemble cast featuring Harry Belafonte, Joy Bryant, Nick Cannon, Laurence Fishburne, Spencer Garrett, Helen Hunt, Anthony Hopkins, Ashton Kutcher, Shia LaBeouf, Lindsay Lohan, William H.Macy, Demi Moore, Martin Sheen, Christian Slater, Sharon Stone, Freddy Rodriguez, Heather Graham, Elijah Wood and Estevez himself. The screenplay is a fictionalized account of the hours leading up to the June 5, 1968 shooting of U.S. Senator Robert F. Kennedy in the kitchen of the Ambassador Hotel in Los Angeles following his win of the 1968 Democratic presidential primary in California. | ✓ | ✗ | ✓ | ✓ | ✗ |
The song was released on November 24, 2014 on iTunes. The song was also released on Spotify, make it available to stream online. One day before the release date, Lil Wayne premiered the single at 42nd American Music Awards ceremony with Christina Milian. The first “Smile" performance was in Family Channel special Family Day, in Canada, on November 16, 2014. On November 23, the band performed at the American Music Awards. On November 28, they performed in Good Morning America. | ✓ | ✗ | ✓ | ✗ | ✗ |
“Quédate esta noche" is a song recorded by Spanish group Trigo Limpio. The song was written by José Antonio Martín. It is best known as the Spanish entry at the Eurovision Song Contest 1980, in The Hague. “Boom Bang-a-Bang" is a song recorded by British singer Lulu. The song was written by Alan Moorhouse and Peter Warne. It is best known as the British winning entry at the Eurovision Song Contest 1969, held in Madrid. | ✗ | ✗ | ✓ | ✓ | ✗ |
Along with von Einem, he is considered to be the best suspect for the Beaumont children abduction as he bore a striking similarity to an identikit picture of the suspect for both the Beaumont children and Adelaide Oval cases. A search for a connection to the Beaumonts was unsuccessful as no employment records existed that could shed light on his movements at the time. Some of the records were believed lost in the 1974 Brisbane flood and it is also possible that Brown, who had unrestricted access to government buildings, may have deleted his files. During World War II he served in the Royal Australian Air Force as a navigator and intelligence officer. Returning to Queensland, he farmed sugarcane and pineapples and joined the Liberal Party. In 1963 he was elected to the Queensland Legislative Assembly as the member for Mount Coot-tha. On 3 October 1975, he was awarded the Queen’s Gallantry Medal for his efforts to rescue a soldier during the flooding in Brisbane the previous year. In 1975 he was appointed to the front bench as Minister for Survey, Valuation, Urban and Regional Affairs, with a further promotion to Attorney-General and Minister for Justice in 1976. | ✓ | ✗ | ✓ | ✓ | ✓ |
In September 2017, New York’s 25th Assembly District Representative Nily Rozic, a Democrat, suggested renaming the park in honor of Heather Heyer, who died in the Charlottesville car attack protesting the Unite the Right rally in Charlottesville, Virginia. No renaming of the park has been undertaken and the proposal was largely ignored. Rozic, along with New York State Senator Brad Hoylman, reintroduced legislation to rename the park in 2018 but it did not make it out of committee. They again introduced legislation in 2019 to rename the park. After covering a Ku Klux Klan rally in Charlottesville, Virginia, she received permission to meet with Jeff Schoep, the leader of the National Socialist Movement. Afterwards, she receives permission to film the group at the Unite the Right rally in Charlottesville, where the group gets into an altercation with and are pepper-sprayed by Antifa counter-protestors. After the death of Heather Heyer and President Donald Trump’s controversial remarks on the rally, Schoep takes Deeyah to the urban decay in Detroit and explains that he moved the organization’s headquarters to the city to take advantage of its economic decline for recruiting. | ✓ | ✗ | ✗ | ✗ | ✗ |
Appendix C Appendix: Other experiments
C.1. Ablation Experiments on Different Rhetorical Relations
To more deeply explore the impact of different rhetorical relations on the effectiveness of coreference models, we conducted ablation experiments on the following rhetorical relations: Background, Elaboration, Attribution, Same-Unit, Joint, Explanation, Enablement, Cause, Topic-Comment, Contrast, Condition, Comparison, Evaluation, Manner-Means, Summary, Temporal, Topic-Change, Textual-Organization.The experi- mental results are shown in Table 13.
The specific approach was to remove the edges corresponding to these rhetorical relations in the constructed graph and add self-loop edges to the corresponding nodes. The experimental results show that for the WEC-Eng test set, the removal of Explanation, Evaluation, Background, and Cause led to the most significant decrease in model performance, with conll f1 scores dropping by 0.4, 0.4, 0.3, and 0.3, respectively. The conll f1 decreases for Temporal, Elaboration, and Enablement were around 0.2, while the remaining rhetorical relations had a minimal impact on the decrease in model performance. For the WEC-Zh test set, the rhetorical relations of Elaboration and Explanation have the most significant impact on the model, with conll f1 scores decreasing by 1.4 and 1.3 respectively. Additionally, the conll f1 scores for Contrast, Summary, and Comparison decreased by 0.8, 0.7, and 0.5 respectively. The impact of the remaining rhetorical relations on the model was relatively minor. This indicates that Elaboration and Explanation play critical roles in the performance of the model on the WEC-Zh test set, highlighting their importance in understanding and modeling discourse. The results also suggest areas for potential improvement in the model, particularly in handling these specific rhetorical relations more effectively.